March, 2011


31
Mar 11

Juntando puntitos en Foursquare

Hace unos días instalé la aplicación de Foursquare en mi teléfono y empecé a usarla. No porque me pareciera realmente buena la idea del sitio sino solamente por la mecánica de juegos aplicada a él. Para los que no lo conocen: cuando uno va a algún lugar, hace un “check-in” en el sitio de Foursquare diciendo que está ahí y puede dejar algún comentario relacionado. Si el lugar es caro, si tiene buena atención, etc. Hasta acá es un gran sitio con tips sobre bares / restaurantes / balnearios / negocios / etc. Pero ¿quién querría hacer todo ese trabajo (y gratis) del check-in y agregar comentarios cuando llega a un lugar? ¿Qué gana?

Continue reading →


30
Mar 11

Hanging indents in ePub

This is meant to complete the first post about indents. Today I’ll talk about hanging indents.

Continue reading →


28
Mar 11

Cover images in ePub

All the book covers I’ve made/seen for ePubs share the same characteristics: The cover is an image, centered horizontally and it resizes dinamically according to the screen resolution.

Continue reading →


27
Mar 11

Text indent in ePubs

This is (hopefully) the first of a series of posts related to ePub production. They’ll be small recipes to solve problems/tasks I’ve faced when producing ePubs. This first one is about text indent.

Continue reading →


24
Mar 11

Harvesting Facebook Directory (revisited)

Last year, Ronald Bowes talked about the privacy issues of Facebook Directory, which has links to all the public Facebook profiles. He scraped the contents of the directory and published the results (his post is Return of the Facebook Snatchers).

The post made me think, not about the privacy issues, but on how much information was there to process. How much time should take to scrape the Directory? How to store the urls?. Ron provided a Ruby script which reads a text file, with one url per line, the script downloads the document and extracts links to other pages on the directory. But’s there’s still missing some other script to iterate over these results and call the first script the continue downloading. And there’s still missing some code to know the first and last name of the user for certain url. So I decided to write my own scraper for Facebook Directory.

Continue reading →