storytelling with data
by Cole Nussbaumer Knaflic
π 284 pages
π Read Online
#datascience #datavisualization
ββββββββββββββ
πJoin @datascience_bds for moreπ
by Cole Nussbaumer Knaflic
π 284 pages
π Read Online
#datascience #datavisualization
ββββββββββββββ
πJoin @datascience_bds for moreπ
Forwarded from AI Revolution
Leap Learning
LEAP by Thoughtjumper is an intelligent learning tool designed to enhance the learning experience. It aims to guide individuals in effective learning across various domains such as business, data science, technology, design, and more.
The tool offers learning quests in a wide range of subjects, allowing users to select their desired topics such as web development, digital marketing, data science, finance, and more.LEAP is focused on helping users learn faster and better.
It provides an intelligent guidance system that adapts to individual learning preferences. By decluttering distractions, LEAP allows users to solely focus on their learning, leading to a more immersive experience.
π°Price: Free
π Link
LEAP by Thoughtjumper is an intelligent learning tool designed to enhance the learning experience. It aims to guide individuals in effective learning across various domains such as business, data science, technology, design, and more.
The tool offers learning quests in a wide range of subjects, allowing users to select their desired topics such as web development, digital marketing, data science, finance, and more.LEAP is focused on helping users learn faster and better.
It provides an intelligent guidance system that adapts to individual learning preferences. By decluttering distractions, LEAP allows users to solely focus on their learning, leading to a more immersive experience.
π°Price: Free
π Link
9 types of data visualization
In this article, I will guide you through the wonderful world of data visualization and expand your knowledge about the way you can display your data and how to tell your data story to your specific audience.
Letβs start with data visualization in its most basic form; the (static) chart. Charts are used to display large amounts of data in a condensed and easy-to-understand manner. They are graphical representations of data which makes it easy and fast to digest by the brain. Moreover, charts make it apparent to find hidden information and insights that are otherwise hard to find from a table with data.
There are a lot of types of charts, each with its own function. The most commonly known charts are the bar chart, the line chart, and the pie chart. Charts form the basis for all types of data visualizations I will discuss in this blog.
π Read More
In this article, I will guide you through the wonderful world of data visualization and expand your knowledge about the way you can display your data and how to tell your data story to your specific audience.
Letβs start with data visualization in its most basic form; the (static) chart. Charts are used to display large amounts of data in a condensed and easy-to-understand manner. They are graphical representations of data which makes it easy and fast to digest by the brain. Moreover, charts make it apparent to find hidden information and insights that are otherwise hard to find from a table with data.
There are a lot of types of charts, each with its own function. The most commonly known charts are the bar chart, the line chart, and the pie chart. Charts form the basis for all types of data visualizations I will discuss in this blog.
π Read More
Ocean Data in Canada
Learn what ocean data are, how they're being used, and the ways in which you can access open ocean data.
Rating βοΈ: 4.7 out 5
Students π¨βπ : 1368
Duration β° : 49min of on-demand video
Created by π¨βπ«: Katherine Luber, Jacob Thompson, Shayla Fitzsimmons
π Course Link
#datascience
ββββββββββββββ
πJoin @datascience_bds for moreπ
Learn what ocean data are, how they're being used, and the ways in which you can access open ocean data.
Rating βοΈ: 4.7 out 5
Students π¨βπ : 1368
Duration β° : 49min of on-demand video
Created by π¨βπ«: Katherine Luber, Jacob Thompson, Shayla Fitzsimmons
π Course Link
#datascience
ββββββββββββββ
πJoin @datascience_bds for moreπ
Udemy
Free Data Science Tutorial - Ocean Data in Canada
Learn what ocean data are, how they're being used, and the ways in which you can access open ocean data. - Free Course
Latex Cheat Sheet of data sceince.pdf
1.4 MB
Latex Cheat Sheet of data science
Your Ultimate guide to Permutations
Have you ever marveled at how many ways you can arrange a set of items when the order truly matters? In this article, I will explain permutations, exploring how they help determine the number of possible arrangements in a set.
If you find my articles interesting, donβt forget to clap and follow ππΌ, these articles take times and effort to do!
Permutations
βA permutation is a mathematical technique that determines the number of possible arrangements in a set when the order of the arrangements matters. Common mathematical problems involve choosing only several items from a set of items in a certain order. β[1]
Types of permutations
1 / Permutations Without Repetition : used when each item in the set can only appear once in each arrangement.
π Read More
Have you ever marveled at how many ways you can arrange a set of items when the order truly matters? In this article, I will explain permutations, exploring how they help determine the number of possible arrangements in a set.
If you find my articles interesting, donβt forget to clap and follow ππΌ, these articles take times and effort to do!
Permutations
βA permutation is a mathematical technique that determines the number of possible arrangements in a set when the order of the arrangements matters. Common mathematical problems involve choosing only several items from a set of items in a certain order. β[1]
Types of permutations
1 / Permutations Without Repetition : used when each item in the set can only appear once in each arrangement.
π Read More
Medium
Your Ultimate guide to Permutations
We are going to cover today a branch of mathematics βCombiatoricsβ, precisely permutations as well as factorial function.
Data Science Core Concepts 2023
Data Science Core Concepts
Rating βοΈ: 4.8 out 5
Students π¨βπ : 1551
Duration β° : 1hr 49min of on-demand video
Created by π¨βπ«: Python Only Geeks
π Course Link
#datascience
ββββββββββββββ
πJoin @datascience_bds for moreπ
Data Science Core Concepts
Rating βοΈ: 4.8 out 5
Students π¨βπ : 1551
Duration β° : 1hr 49min of on-demand video
Created by π¨βπ«: Python Only Geeks
π Course Link
#datascience
ββββββββββββββ
πJoin @datascience_bds for moreπ
Udemy
Free Data Science Tutorial - Data Science Core Concepts 2023
Data Science Core Concepts - Free Course
Ray
Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
Creator: ray-project
Stars βοΈ: 33.3k
Forked By: 5.6k
https://github.com/ray-project/ray
#datascience
ββββββββββββββ
Join @datascience_bds for more cool repositories.
*This channel belongs to @bigdataspecialist group
Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
Creator: ray-project
Stars βοΈ: 33.3k
Forked By: 5.6k
https://github.com/ray-project/ray
#datascience
ββββββββββββββ
Join @datascience_bds for more cool repositories.
*This channel belongs to @bigdataspecialist group
GitHub
GitHub - ray-project/ray: Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries forβ¦
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads. - ray-project/ray
Mastering Probability and Combinatorics
"Mastering the Essentials: Probability and Combinatorics Explained"
Rating βοΈ: 4.0 out 5
Students π¨βπ : 1,129
Duration β° : 1hr 24min of on-demand video
Created by π¨βπ«: Akhil Vydyula
π Course Link
#probability
ββββββββββββββ
πJoin @datascience_bds for moreπ
"Mastering the Essentials: Probability and Combinatorics Explained"
Rating βοΈ: 4.0 out 5
Students π¨βπ : 1,129
Duration β° : 1hr 24min of on-demand video
Created by π¨βπ«: Akhil Vydyula
π Course Link
#probability
ββββββββββββββ
πJoin @datascience_bds for moreπ
Udemy
Free Data Science Tutorial - Mastering Probability and Combinatorics
"Mastering the Essentials: Probability and Combinatorics Explained" - Free Course
Data Science Portfolios, Speeding Up Python, KANs, and Other May Must-Reads
Python One Billion Row Challenge β From 10 Minutes to 4 Seconds
With a longstanding reputation for slowness, youβd think that Python wouldnβt stand a chance at doing well in the popular βone billion rowβ challenge. Dario RadeΔiΔβs viral post aims to show that with some flexibility and outside-the-box thinking, you can still squeeze impressive time savings out of your code.
N-BEATS β The First Interpretable Deep Learning Model That Worked for Time Series Forecasting
Anyone who enjoys a thorough look into a modelβs inner workings should bookmark Jonte Danckerβs excellent explainer on N-BEATS, the βfirst pure deep learning approach that outperformed well-established statistical approachesβ for time-series forecasting tasks.
Build a Data Science Portfolio Website with ChatGPT: Complete Tutorial
In a competitive job market, data scientists canβt afford to be coy about their achievements and expertise. A portfolio website can be a powerful way to showcase both, and Natassha Selvarajβs patient guide demonstrates how you can build one from scratch with the help of generative-AI tools.
A Complete Guide to BERT with Code
Why not take a step back from the latest buzzy model to learn about those precursors that made todayβs innovations possible? Bradney Smith invites us to go all the way back to 2018 (or several decades ago, in AI time) to gain a deep understanding of the groundbreaking BERT (Bidirectional Encoder Representations from Transformers) model.
Why LLMs Are Not Good for Coding β Part II
Back in the present day, we keep hearing about the imminent obsolescence of programmers as LLMs continue to improve. Andrea Valenzuelaβs latest article serves as a helpful βnot so fast!β interjection, as she focuses on their inherent limitations when it comes to staying up-to-date with the latest libraries and code functionalities.
PCA & K-Means for Traffic Data in Python
What better way to round out our monthly selection than with a hands-on tutorial on a core data science workflow? In her debut TDS post, Beth Ou Yang walks us through a real-world exampleβtraffic data from Taiwan, in this caseβof using principle component analysis (PCA) and K-means clustering.
Python One Billion Row Challenge β From 10 Minutes to 4 Seconds
With a longstanding reputation for slowness, youβd think that Python wouldnβt stand a chance at doing well in the popular βone billion rowβ challenge. Dario RadeΔiΔβs viral post aims to show that with some flexibility and outside-the-box thinking, you can still squeeze impressive time savings out of your code.
N-BEATS β The First Interpretable Deep Learning Model That Worked for Time Series Forecasting
Anyone who enjoys a thorough look into a modelβs inner workings should bookmark Jonte Danckerβs excellent explainer on N-BEATS, the βfirst pure deep learning approach that outperformed well-established statistical approachesβ for time-series forecasting tasks.
Build a Data Science Portfolio Website with ChatGPT: Complete Tutorial
In a competitive job market, data scientists canβt afford to be coy about their achievements and expertise. A portfolio website can be a powerful way to showcase both, and Natassha Selvarajβs patient guide demonstrates how you can build one from scratch with the help of generative-AI tools.
A Complete Guide to BERT with Code
Why not take a step back from the latest buzzy model to learn about those precursors that made todayβs innovations possible? Bradney Smith invites us to go all the way back to 2018 (or several decades ago, in AI time) to gain a deep understanding of the groundbreaking BERT (Bidirectional Encoder Representations from Transformers) model.
Why LLMs Are Not Good for Coding β Part II
Back in the present day, we keep hearing about the imminent obsolescence of programmers as LLMs continue to improve. Andrea Valenzuelaβs latest article serves as a helpful βnot so fast!β interjection, as she focuses on their inherent limitations when it comes to staying up-to-date with the latest libraries and code functionalities.
PCA & K-Means for Traffic Data in Python
What better way to round out our monthly selection than with a hands-on tutorial on a core data science workflow? In her debut TDS post, Beth Ou Yang walks us through a real-world exampleβtraffic data from Taiwan, in this caseβof using principle component analysis (PCA) and K-means clustering.
12 Fundamental Math Theories Needed to Understand AI
1. Curse of Dimensionality
This phenomenon occurs when analyzing data in high-dimensional spaces. As dimensions increase, the volume of the space grows exponentially, making it challenging for algorithms to identify meaningful patterns due to the sparse nature of the data.
2. Law of Large Numbers
A cornerstone of statistics, this theorem states that as a sample size grows, its mean will converge to the expected value. This principle assures that larger datasets yield more reliable estimates, making it vital for statistical learning methods.
3. Central Limit Theorem
This theorem posits that the distribution of sample means will approach a normal distribution as the sample size increases, regardless of the original distribution. Understanding this concept is crucial for making inferences in machine learning.
4. Bayesβ Theorem
A fundamental concept in probability theory, Bayesβ Theorem explains how to update the probability of your belief based on new evidence. It is the backbone of Bayesian inference methods used in AI.
5. Overfitting and Underfitting
Overfitting occurs when a model learns the noise in training data, while underfitting happens when a model is too simplistic to capture the underlying patterns. Striking the right balance is essential for effective modeling and performance.
6. Gradient Descent
This optimization algorithm is used to minimize the loss function in machine learning models. A solid understanding of gradient descent is key to fine-tuning neural networks and AI models.
7. Information Theory
Concepts like entropy and mutual information are vital for understanding data compression and feature selection in machine learning, helping to improve model efficiency.
8. Markov Decision Processes (MDP)
MDPs are used in reinforcement learning to model decision-making scenarios where outcomes are partly random and partly under the control of a decision-maker. This framework is crucial for developing effective AI agents.
9. Game Theory
Old school AI is based off game theory. This theory provides insights into multi-agent systems and strategic interactions among agents, particularly relevant in reinforcement learning and competitive environments.
10. Statistical Learning Theory
This theory is the foundation of regression, regularization and classification. It addresses the relationship between data and learning algorithms, focusing on the theoretical aspects that govern how models learn from data and make predictions.
11. Hebbian Theory
This theory is the basis of neural networks, βNeurons that fire together, wire togetherβ. Its a biology theory on how learning is done on a cellular level, and as you would have it β Neural Networks are based off this theory.
12. Convolution (Kernel)
Not really a theory and you donβt need to fully understand it, but this is the mathematical process on how masks work in image processing. Convolution matrix is used to combine two matrixes and describes the overlap.
Special thanks to Jiji Veronica Kim for this list.
ββββββββββββββ
Join @datascience_bds for more cool repositories.
*This channel belongs to @bigdataspecialist group
1. Curse of Dimensionality
This phenomenon occurs when analyzing data in high-dimensional spaces. As dimensions increase, the volume of the space grows exponentially, making it challenging for algorithms to identify meaningful patterns due to the sparse nature of the data.
2. Law of Large Numbers
A cornerstone of statistics, this theorem states that as a sample size grows, its mean will converge to the expected value. This principle assures that larger datasets yield more reliable estimates, making it vital for statistical learning methods.
3. Central Limit Theorem
This theorem posits that the distribution of sample means will approach a normal distribution as the sample size increases, regardless of the original distribution. Understanding this concept is crucial for making inferences in machine learning.
4. Bayesβ Theorem
A fundamental concept in probability theory, Bayesβ Theorem explains how to update the probability of your belief based on new evidence. It is the backbone of Bayesian inference methods used in AI.
5. Overfitting and Underfitting
Overfitting occurs when a model learns the noise in training data, while underfitting happens when a model is too simplistic to capture the underlying patterns. Striking the right balance is essential for effective modeling and performance.
6. Gradient Descent
This optimization algorithm is used to minimize the loss function in machine learning models. A solid understanding of gradient descent is key to fine-tuning neural networks and AI models.
7. Information Theory
Concepts like entropy and mutual information are vital for understanding data compression and feature selection in machine learning, helping to improve model efficiency.
8. Markov Decision Processes (MDP)
MDPs are used in reinforcement learning to model decision-making scenarios where outcomes are partly random and partly under the control of a decision-maker. This framework is crucial for developing effective AI agents.
9. Game Theory
Old school AI is based off game theory. This theory provides insights into multi-agent systems and strategic interactions among agents, particularly relevant in reinforcement learning and competitive environments.
10. Statistical Learning Theory
This theory is the foundation of regression, regularization and classification. It addresses the relationship between data and learning algorithms, focusing on the theoretical aspects that govern how models learn from data and make predictions.
11. Hebbian Theory
This theory is the basis of neural networks, βNeurons that fire together, wire togetherβ. Its a biology theory on how learning is done on a cellular level, and as you would have it β Neural Networks are based off this theory.
12. Convolution (Kernel)
Not really a theory and you donβt need to fully understand it, but this is the mathematical process on how masks work in image processing. Convolution matrix is used to combine two matrixes and describes the overlap.
Special thanks to Jiji Veronica Kim for this list.
ββββββββββββββ
Join @datascience_bds for more cool repositories.
*This channel belongs to @bigdataspecialist group
streamlit
Streamlit β A faster way to build and share data apps.
Creator: Streamlit
Stars βοΈ: 35.4k
Forked By: 3.1k
https://github.com/streamlit/streamlit
#datascience
ββββββββββββββ
Join @datascience_bds for more cool repositories.
*This channel belongs to @bigdataspecialist group
Streamlit β A faster way to build and share data apps.
Creator: Streamlit
Stars βοΈ: 35.4k
Forked By: 3.1k
https://github.com/streamlit/streamlit
#datascience
ββββββββββββββ
Join @datascience_bds for more cool repositories.
*This channel belongs to @bigdataspecialist group
GitHub
GitHub - streamlit/streamlit: Streamlit β A faster way to build and share data apps.
Streamlit β A faster way to build and share data apps. - streamlit/streamlit