Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to do Data Science with larger than memory data using Dask? #132

Open
arnabbiswas1 opened this issue Dec 21, 2020 · 3 comments
Open
Assignees
Labels

Comments

@arnabbiswas1
Copy link

  • Abstract (2-3 lines)

As a Data Scientist, we face few challenges while dealing with large volume of data:

  1. Popular Python libraries like NumPy & Pandas are not designed to scale beyond single processor/core
  2. Numpy, Pandas, Scikit-Learn are not designed to scale beyond a single machine
  3. If data is bigger than RAM, these libraries can't be used

In this session, I will discuss how these challenges can be addressed using parallel computing library, Dask.

  • Brief Description and Contents to be covered

The talk is divided in two portions:

  1. Understanding the challenges of large data (Will be delivered through presentation)
    a. Fundamentals of computer architecture (with a focus on Computing Unit & Memory unit)
    b. Why parallelism is necessary in a multi-core architecture?
    c. Challenges with large data (data that doesn't fit RAM) & how to address
    d. Introduction to distributed computing?

  2. How does Dask handle large data? (Code walk through)
    a. What is Dask and Why it is needed?
    b. How Dask parallelizes jobs across cores/processors?
    c. How Dask handles larger than memory data using out of core computing and distributed computing?

  • Pre-requisites for the talk

Basic knowledge about the Python based Data Science libraries like Pandas, NumPy, ScikitLearn

  • Time required for the talk

45 minutes to 1 hour. This talk can be extended to a 2 hour long work shop as well.

  • Link to slides

https://speakerdeck.com/arnabbiswas1/scale-up-your-data-science-work-flow-using-dask

  • Will you be doing hands-on demo as well?

Yes.

  • Link to ipython notebook (if any)

https://github.com/arnabbiswas1/dask_workshop

  • About yourself

https://arnab.blog/about/

  • Are you comfortable if the talk is recorded and uploaded to PyData Delhi's YouTube channel ?

Yes

  • Any query ?

This talk (45 minutes) have been delivered recently to Bangalore Python User Group, BangPypers. Here is the recording for your reference: Link

@MSanKeys963
Copy link
Member

Hi @arnabbiswas1. Thanks for the proposal.
@shagunsodhani please have a look.

@shagunsodhani
Copy link
Contributor

Hey @arnabbiswas1 ! Thanks for proposing the talk. The content looks good. Best of luck :)

@arnabbiswas1
Copy link
Author

@MSanKeys963 @shagunsodhani

As mentioned in the proposal, this talk can be delivered as it is (45 minutes). Or it can be extended to 1.5 to 2 hours focusing on usage of Dask just for Data Science use cases (Existing content + Dask Machine Learning + Dask Dashboard for debugging).

Please let me know your thoughts.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

3 participants