51 avsnitt • Längd: 40 min • Oregelbundet
Machine learning audio course, teaching the fundamentals of machine learning and artificial intelligence. It covers intuition, models (shallow and deep), math, languages, frameworks, etc. Where your other ML resources provide the trees, I provide the forest. Consider MLG your syllabus, with highly-curated resources for each episode’s details at ocdevel.com. Audio is a great supplement during exercise, commute, chores, etc.
The podcast Machine Learning Guide is created by OCDevel. The podcast and the artwork on this page are embedded on this page using the public podcast feed (RSS).
Links
Links
According to the Aider Leaderboard (as of April 12, 2025), leading models include for vibe-coding:
Links
I currently favor Roo Code. Plus either gemini-2.5-pro-exp-03-25 for Architect, Boomerang, or Code with large contexts. And Claude 3.7 for code with small contexts, eg Boomerang subtasks. Many others favor Cursor, Aider, or Cline. Copilot and Windsurf are less vogue lately. I found Copilot to struggle more; and their pricing - previously their winning point - is less compelling now.
Why I favor Roo. The default settings have it as stable and effective as Cline, Cursor. But you can tinker more with these settings - eg, for Gemini 2.5 I disable partial file reads (since it has a huge context window). Their modes are elegantly just custom system prompts (an oversimplification), making custom workflows very powerful. A potent example is their Boomerang Mode, which is an orchestrator that delegates planning and edit subtasks, to keep context windows tight. Boomerang mode specifically is a plugin-seller, it's incredibly powerful. Aider is still a darn decent exacto-knife, but as Roo has grown, I haven't found much need for Aider.
Tools discussed:
Other:
"Vibe coding" using AI agents in software development. It uses LLMs for code generation and project management. Developers are increasingly relying on agentic tools and IDE plugins to improve productivity.
Use of AI in Code GenerationLinks:
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-21
Raybeam and Databricks: Ming Chang from Raybeam discusses Raybeam's focus on data science and analytics, and how their recent acquisition by Dept Agency has expanded their scope into ML Ops and AI. Raybeam often utilizes Databricks due to its comprehensive nature.
Understanding Databricks: Contrary to initial assumptions, Databricks is not just an analytics platform like Tableau but an ML Ops platform competing with tools like SageMaker and Kubeflow. It offers functionalities for creating notebooks, executing Python code, and using a hosted Spark cluster and Delta Lake for data storage.
Choosing the Right MLOps Tool: Depending on client requirements, Raybeam might recommend different tools. Decision factors include client's existing expertise, infrastructure needs, and scaling challenges. Databricks is often recommended for its ease of use and features.
Databricks Features: Offers a hosted solution for Spark clusters on AWS, Azure, or GCP; integrates with IDEs like VSCode through Databricks Connect; provides a unique Git integration for version control of notebooks; and utilizes Delta Lake for version control of Parquet files, enhancing operations like edit and delete.
Parquet and Delta Lake: Parquet files are optimized for big data, and Delta Lake provides transaction-like operations over Parquet by maintaining version history.
Pricing and Usage: Databricks adds a nominal fee on top of cloud provider charges. It's accessible for single developers and startups, making it suitable for various scales of operations.
Ming Chang's Picks: Discusses interests in automated stock trading projects and building drones with Raspberry Pi, highlighting the intersection of programming and physical computing.
For a hands-on look at Ming Chang's drone project, follow his developments or connect for insights on building a Raspberry Pi-powered drone.
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-20
Conversation with Dirk-Jan Kubeflow (vs cloud native solutions like SageMaker)
Dirk-Jan Verdoorn - Data Scientist at Dept Agency
Kubeflow. (From the website:) The Machine Learning Toolkit for Kubernetes. The Kubeflow project is dedicated to making deployments of machine learning (ML) workflows on Kubernetes simple, portable and scalable. Our goal is not to recreate other services, but to provide a straightforward way to deploy best-of-breed open-source systems for ML to diverse infrastructures. Anywhere you are running Kubernetes, you should be able to run Kubeflow.
TensorFlow Extended (TFX). If using TensorFlow with Kubeflow, combine with TFX for maximum power. (From the website:) TensorFlow Extended (TFX) is an end-to-end platform for deploying production ML pipelines. When you're ready to move your models from research to production, use TFX to create and manage a production pipeline.
Alternatives:
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-19
Chatting with co-workers about the role of DevOps in a machine learning engineer's life
Expert coworkers at Dept
Devops tools
Pictures (funny and serious)
Try a walking desk to stay healthy while you study or work!
Show notes: ocdevel.com/mlg/mla-17
Developing on AWS first (SageMaker or other)
Consider developing against AWS as your local development environment, rather than only your cloud deployment environment. Solutions:
Connect to deployed infrastructure via Client VPN
Infrastructure as Code
Try a walking desk to stay healthy while you study or work!
Full note at ocdevel.com/mlg/mla-16
Part 2 of deploying your ML models to the cloud with SageMaker (MLOps)
MLOps is deploying your ML models to the cloud. See MadeWithML for an overview of tooling (also generally a great ML educational run-down.)
Try a walking desk to stay healthy while you study or work!
Show notes Part 1 of deploying your ML models to the cloud with SageMaker (MLOps)
MLOps is deploying your ML models to the cloud. See MadeWithML for an overview of tooling (also generally a great ML educational run-down.)
And I forgot to mention JumpStart, I'll mention next time.
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-14
Server-side ML. Training & hosting for inference, with a goal towards serverless. AWS SageMaker, Batch, Lambda, EFS, Cortex.dev
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-13
Client, server, database, etc.
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-12
Use Docker for env setup on localhost & cloud deployment, instead of pyenv / Anaconda. I recommend Windows for your desktop.
Try a walking desk to stay healthy while you study or work!
Show notes at ocdevel.com/mlg/32.
L1/L2 norm, Manhattan, Euclidean, cosine distances, dot product
Normed distances link
Dot product
Cosine (normalized dot)
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-11
Kmeans (sklearn vs FAISS), finding n_clusters via inertia/silhouette, Agglomorative, DBSCAN/HDBSCAN
Try a walking desk to stay healthy while you study or work!
Full note at ocdevel.com/mlg/mla-10
NLTK: swiss army knife. Gensim: LDA topic modeling, n-grams. spaCy: linguistics. transformers: high-level business NLP tasks.
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-9
matplotlib, Seaborn, Bokeh, D3, Tableau, Power BI, QlikView, Excel
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-8
EDA + charting. DataFrame info/describe, imputing strategies. Useful charts like histograms and correlation matrices.
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-7
Run your code + visualizations in the browser: iPython / Jupyter Notebooks.
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-6
Salary based on location, gender, age, tech... from O'Reilly.
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-5
Dimensions, size, and shape of Numpy ndarrays / TensorFlow tensors, and methods for transforming those.
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-3
Comparison of different data storage options when working with your ML models.
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-2
Some numerical data nitty-gritty in Python.
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/mla-1
Reboot on the MLG episode, with more confident recommends.
Try a walking desk to stay healthy while you study or work!
Introduction to reinforcement learning concepts. ocdevel.com/mlg/29 for notes and resources.
Notes and resources: ocdevel.com/mlg/28
Try a walking desk to stay healthy while you study or work!
More hyperparameters for optimizing neural networks. A focus on regularization, optimizers, feature scaling, and hyperparameter search methods.
Hyperparameter Search TechniquesFull notes and resources at ocdevel.com/mlg/27
Try a walking desk to stay healthy while you study or work!
Hyperparameters are crucial elements in the configuration of machine learning models. Unlike parameters, which are learned by the model during training, hyperparameters are set by humans before the learning process begins. They are the knobs and dials that humans can control to influence the training and performance of machine learning models.
Definition and ImportanceHyperparameters differ from parameters like theta in linear and logistic regression, which are learned weights. They are choices made by humans, such as the type of model, number of neurons in a layer, or the model architecture. These choices can have significant effects on the model's performance, making them vital to conscious and informed tuning.
Types of Hyperparameters Model Selection:Choosing what model to use is itself a hyperparameter. For example, deciding between linear regression, logistic regression, naive Bayes, or neural networks.
Architecture of Neural Networks:They transform linear outputs into non-linear outputs. Popular choices include ReLU, tanh, and sigmoid, with ReLU being the default for most neural network layers.
Regularization and Optimization:These influence the learning process. The use of L1/L2 regularization or dropout, as well as the type of optimizer (e.g., Adam, Adagrad), are hyperparameters.
Optimization TechniquesTechniques like grid search, random search, and Bayesian optimization are used to systematically explore combinations of hyperparameters to find the best configuration for a given task. While these methods can be computationally expensive, they are necessary for achieving optimal model performance.
Challenges and Future DirectionsThe field strives towards simplifying the choice of hyperparameters, ideally automating them to become parameters of the model itself. Efforts like Google's AutoML aim to handle hyperparameter tuning automatically.
Understanding and optimizing hyperparameters is a cornerstone in machine learning, directly impacting the effectiveness and efficiency of a model. Progress continues to integrate these choices into model training, reducing the dependency on human intervention and trial-and-error experimentation.
Decision TreeTry a walking desk to stay healthy while you study or work!
Ful notes and resources at ocdevel.com/mlg/26
NOTE. This episode is no longer relevant, and tforce_btc_trader no longer maintained. The current podcast project is Gnothi.
Episode OverviewThe project "Bitcoin Trader" involves developing a Bitcoin trading bot using machine learning to capitalize on the hot topic of cryptocurrency and its potential profitability. The project will serve as a medium to delve into complex machine learning engineering topics, such as hyperparameter selection and reinforcement learning, over subsequent episodes.
Cryptocurrency, specifically Bitcoin, is used for its universal and decentralized nature, akin to a digital, secure, and democratic financial instrument like the US dollar. Bitcoin mining involves running complex calculations to manage the currency's existence, similar to a distributed Federal Reserve system, with transactions recorded on a secure and permanent ledger known as the blockchain.
The flexibility of cryptocurrency trading allows for machine learning applications across unsupervised, supervised, and reinforcement learning paradigms. This project will focus on using models such as LSTM recurrent neural networks and convolutional neural networks, highlighting Bitcoin’s unique capacity to illustrate machine learning concept decisions like network architecture.
Trading differs from investing by focusing on profit from price fluctuations rather than a belief in long-term value increase. It involves understanding patterns in price actions to buy low and sell high. Different types of trading include day trading, which involves daily buying and selling, and swing trading, which spans longer periods.
Trading decisions rely on patterns identified in price graphs, using time series data. Data representation through candlesticks (OHLCV: open-high-low-close-volume), coupled with indicators like moving averages and RSI, provide multiple input features for machine learning models, enhancing prediction accuracy.
Exchanges like GDAX and Kraken serve as platforms for converting traditional currencies into cryptocurrencies. The efficient market hypothesis suggests that the value of an instrument is fairly priced based on the collective analysis of market participants. Differences in exchange prices can provide opportunities for arbitrage, further fueling trading strategies.
The project code, currently using deep reinforcement learning via tensor force, employs convolutional neural networks over LSTM to adapt to Bitcoin trading's intricacies. The project will be available at ocdevel.com for community engagement, with future episodes tackling hyperparameter selection and deep reinforcement learning techniques.
Try a walking desk to stay healthy while you study or work!
Notes and resources at ocdevel.com/mlg/25
Filters and Feature Maps: Filters are small matrices used to detect visual features from an input image by applying them to local pixel patches, creating a 3D output called a feature map. Each filter is tasked with recognizing a specific pattern (e.g., edges, textures) in the input images.
Convolutional Layers: The filter is applied across the image to produce an output which is the feature map. A convolutional layer is composed of several feature maps, with depth corresponding to the number of filters applied.
Image Compression Techniques:
Max Pooling: Max pooling is a downsampling technique used to reduce the spatial dimensions of feature maps by taking the maximum value over a defined window, further compressing and reducing computational load.
Predefined Architectures: There are well-established predefined architectures like LeNet, AlexNet, and ResNet, which have been fine-tuned through competitions such as the ImageNet Challenge, and can be used directly or adapted for specific tasks in computer vision.
Try a walking desk to stay healthy while you study or work!
Notes and resources at ocdevel.com/mlg/24
HardwareDesktop if you're stationary, as you'll get the best performance bang-for-buck and improved longevity; laptop if you're mobile.
Desktops. Build your own PC, better value than pre-built. See PC Part Picker, make sure to use an Nvidia graphics card. Generally shoot for 2nd-best of CPUs/GPUs. Eg, RTX 4070 currently (2024-01); better value-to-price than 4080+.
For laptops, see this post (updated).
OS / SoftwareUse Linux (I prefer Ubuntu), or Windows, WSL2, and Docker. See mla/12 for details.
Programming Tech StackDeep-learning frameworks. You'll use both TF & PT eventually, so don't get hung up. mlg/9 for details.
Shallow-learning / utilities: ScikitLearn, Pandas, Numpy
Cloud-hosting: AWS / GCP / Azure. mla/13 for details.
Episode SummaryThe episode discusses setting up a tech stack tailored for machine learning, emphasizing the necessity of choosing a primary programming language and framework, which, in this case, are Python and TensorFlow. The decision is supported by the ongoing popularity and community support for these tools. This preference is further influenced by the necessity for GPU optimization, which TensorFlow provides, allowing for enhanced performance through utilizing Nvidia's CUDA technology.
A notable change in the landscape is the decline of certain deep learning frameworks such as Theano, and the rise of competitors like PyTorch, which is gaining traction due to its ease of use in comparison to TensorFlow. The author emphasizes the importance of selecting frameworks with robust community support and resources, highlighting TensorFlow's lead in the market in this respect.
For hardware, the suggestion is a custom-built PC with a powerful Nvidia GPU, such as the 1080 TI, running Ubuntu Linux for best compatibility. However, for those who favor cloud services, Amazon Web Services (AWS) and Google Cloud Platform (GCP) are viable options, with a preference for GCP due to cost and performance benefits, particularly with the upcoming Tensor Processing Units (TPUs).
On the software side, the use of Pandas for data manipulation, NumPy for mathematical operations, and Scikit-Learn for shallow learning tasks provides a comprehensive toolkit for machine learning development. Additionally, the use of abstraction libraries such as Keras for simplifying TensorFlow syntax and TensorForce for reinforcement learning are recommended.
The episode further explores system architectures, suggesting a separation of concerns between a web app server and a machine learning (job) server. Communication between these components can be efficiently managed using a message queuing system like RabbitMQ, with Celery as a potential abstraction layer.
To support developers in implementing their machine learning pipelines, the recommendation extends to leveraging existing datasets, using Scikit-Learn for convenient access, and standardizing data for effective training results. The author points to several books and resources to assist in understanding and applying these technologies effectively, ending with your own workstation recommendations and building TensorFlow from source for performance gains as a potential advanced optimization step.
Try a walking desk to stay healthy while you study or work!
Notes and resources at ocdevel.com/mlg/23
Neural Network Types in NLPVanilla Neural Networks (Feedforward Networks):
Convolutional Neural Networks (CNNs):
Recurrent Neural Networks (RNNs):
Supervised vs Reinforcement Learning:
Encoder-Decoder Models:
Gradient Problems & Solutions:
Try a walking desk to stay healthy while you study or work!
Notes and resources at ocdevel.com/mlg/22
Deep NLP Fundamentals
Deep learning has had a profound impact on natural language processing by introducing models like recurrent neural networks (RNNs) that are specifically adept at handling sequential data. Unlike traditional linear models like linear regression, RNNs can address the complexities of language which appear from its inherent non-linearity and hierarchy. These models are able to learn complex features by combining data in multiple layers, which has revolutionized areas like sentiment analysis, machine translation, and more.
Neural Networks and Their Use in NLP
Neural networks can be categorized into regular feedforward neural networks and recurrent neural networks (RNNs). Feedforward networks are used for non-sequential tasks, while RNNs are useful for sequential data processing such as language, where the network’s hidden layers are connected to enable learning over time steps. This loopy architecture allows RNNs to maintain a form of state or memory, making them effective for tasks where context is crucial. The challenge of mapping these sequences into meaningful output has led to architectures like the encoder-decoder model, which reads entire sequences to produce responses or translations, enhancing the network's ability to learn and remember context across long sequences.
Word Embeddings and Contextual Representations
A key challenge in processing natural language using machine learning models is representing words as numbers, as machine learning relies on mathematical operations. Initial representations like one-hot vectors were simple but lacked semantic meaning. To address this, word embeddings such as those generated by the Word2Vec model have been developed. These embeddings place words in a vector space where distance and direction between vectors are meaningful, allowing models to interpret semantic similarities and differences between words. Word2Vec, using neural networks, learns these embeddings by predicting word contexts or vice versa.
Advanced Architectures and Practical Implications
RNNs and their more sophisticated versions like LSTM and GRU cells address specific challenges such as the vanishing gradient problem, which can occur during backpropagation through time. These architectures allow for more effective and longer-range dependencies to be learned, vital for handling the nuances of human language. As a result, these models have become dominant in modern NLP, replacing older methods for tasks ranging from part-of-speech tagging to machine translation.
Further Learning and Resources
For in-depth learning, resources such as the "Unreasonable Effectiveness of RNNs", Stanford courses on deep NLP by Christopher Manning, and continued education in deep learning can enhance one's understanding of these models. Emphasis on both theoretical understanding and practical application will be crucial for mastering the deep learning techniques that are transforming NLP.
Try a walking desk to stay healthy while you study or work!
Notes and resources at ocdevel.com/mlg/20
NLP progresses through three main layers: text preprocessing, syntax tools, and high-level goals, each building upon the last to achieve complex linguistic tasks.
Text PreprocessingText preprocessing involves essential steps such as tokenization, stemming, and stop word removal. These foundational tasks clean and prepare text for further analysis, ensuring that subsequent processes can be applied more effectively.
Syntax ToolsSyntax tools are crucial for understanding grammatical structures within text. Part of Speech Tagging identifies the role of words within sentences, such as noun, verb, or adjective. Named Entity Recognition (NER) distinguishes entities such as people, organizations, and dates, leveraging models like maximum entropy, support vector machines, or hidden Markov models.
Achieving High-Level GoalsHigh-level NLP goals include text classification, sentiment analysis, and optimizing search engines. Techniques such as the Naive Bayes algorithm enable effective text classification by simplifying documents into word occurrence models. Search engines benefit from the TF-IDF method in tandem with cosine similarity, allowing for efficient document retrieval and relevance ranking.
In-depth Look at Syntax ParsingSyntax parsing delves into sentence structure through two primary approaches: context-free grammars (CFG) and dependency parsing. CFGs use production rules to break down sentences into components like noun phrases and verb phrases. Probabilistic enhancements to CFGs learn from datasets like the Penn Treebank to determine the likelihood of various grammatical structures. Dependency parsing, on the other hand, maps out word relationships through directional arcs, providing a visual dependency tree that highlights connections between components such as subjects and verbs.
Applications of NLP ToolsSyntax parsing plays a vital role in tasks like relationship extraction, providing insights into how entities relate within text. Question answering integrates various tools, using TF-IDF and syntax parsing to locate and extract precise answers from relevant documents, evidenced in systems like Google’s snippet answers.
Text summarization seeks to distill large texts into concise summaries. By employing TF-IDF, the process identifies sentences rich in informational content due to their less frequent vocabulary, removing redundancies for a coherent summary. TextRank, a graph-based methodology, evaluates sentence importance based on their connectedness within a document.
Machine Translation EvolutionMachine translation demonstrates the transformative impact of deep learning. Traditional methods, characterized by their complexity and multiple models, have been surpassed by neural machine translation systems. These employ recurrent neural networks (RNNs) to achieve end-to-end translation, accommodating tasks traditionally dependent on separate linguistic models into a unified approach, thus simplifying development and improving accuracy.
The episode underscores the transition from shallow NLP approaches to deep learning methods, highlighting how advanced models, particularly those involving RNNs, are redefining speech processing tasks with efficiency and sophistication.
Try a walking desk to stay healthy while you study or work!
Notes and resources at ocdevel.com/mlg/19
Classical NLP Techniques:
Origins and Phases in NLP History: Initially reliant on hardcoded linguistic rules, NLP's evolution significantly pivoted with the introduction of machine learning, particularly shallow learning algorithms, leading eventually to deep learning, which is the current standard.
Importance of Classical Methods: Knowing traditional methods is still valuable, providing a historical context and foundation for understanding NLP tasks. Traditional methods can be advantageous with small datasets or limited compute power.
Edit Distance and Stemming:
Language Models:
Naive Bayes for Classification:
Part of Speech Tagging and Named Entity Recognition:
Generative vs. Discriminative Models:
Topic Modeling with LDA:
Search and Similarity Measures:
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/18
Overview: Natural Language Processing (NLP) is a subfield of machine learning that focuses on enabling computers to understand, interpret, and generate human language. It is a complex field that combines linguistics, computer science, and AI to process and analyze large amounts of natural language data.
NLP StructureNLP is divided into three main tiers: parts, tasks, and goals.
1. PartsText Pre-processing:
Syntactic Analysis:
High-Level Applications:
Evolution:
Key Algorithms:
NLP offers robust career prospects as companies strive to implement technologies like chatbots, virtual assistants (e.g., Siri, Google Assistant), and personalized search experiences. It's integral to market leaders like Google, which relies on NLP for applications from search result ranking to understanding spoken queries.
Resources for Learning NLPBooks:
Online Courses:
Tools and Libraries:
NLP continues to evolve with applications expanding across AI, requiring collaboration with fields like speech processing and image recognition for tasks like OCR and contextual text understanding.
Try a walking desk to stay healthy while you study or work!
At this point, browse #importance:essential on ocdevel.com/mlg/resources with the 45m/d ML, 15m/d Math breakdown.
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/16
Inspiration in AI DevelopmentEarly inspirations for AI development centered around solving challenging problems, but recent advancements like self-driving cars and automated scientific discoveries attract professionals due to potential economic automation and career opportunities.
The SingularityThe singularity suggests exponential technological growth leading to a point where AI and robotics automate all technology development, potentially achieving 'seed AI' capable of self-improvement and escaping human intervention.
Defining ConsciousnessConsciousness distinguishes intelligence by awareness. Perception, self-identity, learning, memory, and awareness might all contribute to consciousness, but awareness or subjective experience (quaia) is viewed as a core component.
Hard vs. Soft Problems of ConsciousnessThe soft problems are those we know through sciences — like brain regions being associated with specific functions. The hard problem, however, is explaining how subjective experience arises from physical processes in the brain.
Theories and DebatesOpinions vary widely on whether AI can achieve consciousness, depending on theories around biological plausibility and arguments like John Searl's Chinese Room. The matter of consciousness remains deeply philosophical, touching on human identity itself. The expansion of machine learning and AI might be humanity's next evolutionary step, potentially culminating in the creation of conscious entities.
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/15
ConceptsTry a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/14
Anomaly Detection SystemsTry a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/13
Support Vector Machines (SVM)Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/12
TopicsShallow vs. Deep Learning: Shallow learning can often solve problems more efficiently in time and resources compared to deep learning.
Supervised Learning: Key algorithms include linear regression, logistic regression, neural networks, and K Nearest Neighbors (KNN). KNN is unique as it is instance-based and simple, categorizing new data based on proximity to known data points.
Unsupervised Learning:
Decision Trees: Utilized for both classification and regression, decision trees offer a visible, understandable model structure. Variants like Random Forests and Gradient Boosting Trees increase performance and reduce overfitting risks.
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/10
Topics:Recommended Languages and Frameworks:
Language Choices:
Framework Details:
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/9
Key Concepts:
Unique Features of Neural Networks:
Applications:
Computational Considerations:
Architectures & Optimization:
Try a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/8
Mathematics in Machine LearningTry a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/7. See Andrew Ng Week 3 Lecture Notes
OverviewTry a walking desk to stay healthy while you study or work!
Full notes at ocdevel.com/mlg/6
Pursuing Machine Learning:Try a walking desk to stay healthy while you study or work!
Show notes at ocdevel.com/mlg/5. See Andrew Ng Week 2 Lecture Notes
Key ConceptsAccess to Andrew Ng's Course on Coursera is encouraged to gain in-depth understanding and application skills in machine learning.
Try a walking desk to stay healthy while you study or work!
Show notes at ocdevel.com/mlg/4
The AI HierarchyTry a walking desk to stay healthy while you study or work!
Show notes at ocdevel.com/mlg/3.
This episode covers four major philosophical topics related to artificial intelligence. The purpose is to give broader context to why AI matters, before moving into technical details in later episodes.
1. Economic AutomationAI is automating not just simple tasks like data entry or tax prep, but also high-skill jobs such as medical diagnostics, surgery, and creative work like design, music, and art. There are two common reactions:
The singularity refers to a point of runaway technological growth, where AI becomes capable of improving itself recursively. This concept is tied to "artificial general intelligence" and "seed AI"—systems that not only perform tasks but create better versions of themselves. The idea is that this could trigger extremely rapid change, possibly representing a new phase of evolution beyond humanity.
3. ConsciousnessI explore whether consciousness can emerge from machines. Since the brain is a physical machine and consciousness arises from it, it's possible that artificial systems could develop similar properties. Related ideas:
I discuss scenarios where AI causes harm not through malevolence but through poorly defined objectives. One example is the "paperclip maximizer" thought experiment, where an AI tasked with maximizing paperclip production might consume all resources to do so. This has led some public figures to raise concerns about AI safety. I don't share the same level of concern, but the topic is worth being aware of.
ReferencesIn the next episode, I begin covering the technical foundations of machine learning, starting with supervised, unsupervised, and reinforcement learning.
Links:
What is artificial intelligence, machine learning, and data science? What are their differences? AI history.
Hierarchical breakdown: DS(AI(ML)). Data science: any profession dealing with data (including AI & ML). Artificial intelligence is simulated intellectual tasks. Machine Learning is algorithms trained on data to learn patterns to make predictions.
Artificial Intelligence (AI) - WikipediaOxford Languages: the theory and development of computer systems able to perform tasks that normally require human intelligence, such as visual perception, speech recognition, decision-making, and translation between languages.
AlphaGo Movie, very good!
Applications
Oxford Languages: the use and development of computer systems that are able to learn and adapt without following explicit instructions, by using algorithms and statistical models to analyze and draw inferences from patterns in data.
Data Science (DS) - WikipediaWikipedia: Data science is an interdisciplinary field that uses scientific methods, processes, algorithms and systems to extract knowledge and insights from noisy, structured and unstructured data, and apply knowledge and actionable insights from data across a broad range of application domains. Data science is related to data mining, machine learning and big data.
History1700s-1800s: Statistics & Mathematical decision making
1936: Universal Turing Machine
50s-70s: "AI" coined @Dartmouth workshop 1956 - goal to simulate all aspects of intelligence. John McCarthy, Marvin Minksy, Arthur Samuel, Oliver Selfridge, Ray Solomonoff, Allen Newell, Herbert Simon
90s: Data, Computation, Practical Application -> AI back (90s)
Show notes: ocdevel.com/mlg/1. MLG teaches the fundamentals of machine learning and artificial intelligence. It covers intuition, models, math, languages, frameworks, etc. Where your other ML resources provide the trees, I provide the forest. Consider MLG your syllabus, with highly-curated resources for each episode's details at ocdevel.com. Audio is a great supplement during exercise, commute, chores, etc.
Who is it for
Why audio?
What it's not
Planned episodes
En liten tjänst av I'm With Friends. Finns även på engelska.