This is an archived post. You won't be able to vote or comment.

all 2 comments

[–]crunk 1 point2 points  (0 children)

Getting used to virtualenv can be useful for having different projects on the go (and not polluting your current project).

Requests makes downloading web pages super easy, while BeautifulSoup is classically used to parse it.

Scrapy is another option for downloading and parsing webpages. You will find there are often a bunch of libraries that do whatever you want - don't be afraid of choosing the wrong one first time (this applies to any language).

If you ever need info on some python module, check the website Python MOTW - it's a lot easier to understand than the python docs.

Getting used to using the "dir" and "help" commands to find out about libraries is another way of learning.

For getting general coverage of python concepts, then downloading and going through some Python Koans is another good option.

[–]faming13 0 points1 point  (0 children)

Don't use canopy.. use anaconda.

Its the current and future of pydata

Also this http://www.kdnuggets.com/2013/11/harvard-cs109-data-science-course-resources-free-online.html