Tales of Science & Data
Meta
About me
GitHub
Twitter
Search…
⌃K
Links
Tales of Science and Data
Meta & resources
The meta on all this
Beautiful web of data science
Probability, statistics and data analysis
Probability, its interpretation, and statistics
Foundational concepts on distribution and measures
Hypothesis testing
Methods, theorems & laws
Notable brain teasers, paradoxes and how to be careful with data
Machine Learning: concepts & procedures
Overview of the field
Learning algorithms
Feature building and modelling techniques
Dimensionality reduction and matrix factorisation
Machine Learning: fundamental algorithms
Learning paradigms
Supervised learning
Unsupervised learning
Machine Learning: model assessment
Generic problems models can have
Performance metrics and validation techniques
Training, validation and test sets
Classification metrics
Regression metrics
Cross-validation
Diagnostics
Artificial neural networks
Overview of neural networks
Types of neurons and networks
Natural language processing
General concepts & tasks in NLP
Manipulating text and extracting information
Topic Modelling
Word Embeddings
Computer vision
Intro: quantifying images & some glossary
Processing an image
What's in an image
The computer science appendix
What's this
Notes on foundations
Essential algorithms
The mathematics appendix
Matrix algebra notes
Mathematical functions
Some geometry
Cross-field concepts
(Some) mathematical measures
Toolbox
The Python data stack
Databases and distributed frameworks
Notebook tools
Powered By
GitBook
Performance metrics and validation techniques
The metrics used to assess the quality of a model.
Code & notebooks
Code in this section can also be seen in the repo notebooks
here
.
Contents
Training, validation and test sets
Classification metrics
Regression metrics
Cross-validation
Previous
Overfitting and underfitting
Next
Training, validation and test sets
Last modified
2yr ago