Skip to content

Tags

Significance

Intuitive summaries

Definitions

  • A system that is intelligent and constructed by humans.
  • A branch of computer science which develops and studies intelligent machines.
  • Artificial intelligence (AI), in its broadest sense, is intelligence exhibited by machines, particularly computer systems. It is a field of research in computer science that develops and studies methods and software which enable machines to perceive their environment and uses learning and intelligence to take actions that maximize their chances of achieving defined [[goal|goals]].

Technical summaries

Main resources

Landscapes

Crossovers

Resources

Stanford machine learning https://www.youtube.com/playlist?list=PLoROMvodv4rNyWOpJg_Yh4NSqI4Z4vOYy
Stanford machine learning https://www.youtube.com/playlist?list=PLoROMvodv4rMiGQp3WXShtMGgzqpfVfbU
Stanford transformers https://www.youtube.com/playlist?list=PLoROMvodv4rNiJRchCzutFw5ItR_Z27CM
Stanford generative models including diffusion https://www.youtube.com/playlist?list=PLoROMvodv4rPOWA-omMM6STXaWW4FvJT8
Stanford deep learning https://www.youtube.com/playlist?list=PLoROMvodv4rOABXSygHTsbvUz4G_YQhOb
Karpathy neural networks zero to hero https://www.youtube.com/playlist?list=PLAqhIrjkxbuWI23v9cThsA9GvCAUhRvKZ Stanford natural language processing with deep learning https://www.youtube.com/playlist?list=PLoROMvodv4rMFqRtEuo6SGjY4XbRIVRd4
MIT deep learning https://www.youtube.com/playlist?list=PLTZ1bhP8GBuTCqeY19TxhHyrwFiot42_U
Stanford artificial intelligence https://www.youtube.com/playlist?list=PLoROMvodv4rO1NB9TD4iUZ3qghGEGtqNX
Stanford machine learning with graphs https://www.youtube.com/playlist?list=PLoROMvodv4rPLKxIpqhjhPgdQy7imNkDn
Stanford natural language understanding https://www.youtube.com/playlist?list=PLoROMvodv4rOwvldxftJTmoR3kRcWkJBp
Stanford reinforcement learning https://www.youtube.com/playlist?list=PLoROMvodv4rOSOPzutgyCTapiGlY2Nd8u
Stanford meta-learning https://www.youtube.com/playlist?list=PLoROMvodv4rNjRoawgt72BBNwL2V7doGI
Stanford artificial intelligence https://www.youtube.com/playlist?list=PLoROMvodv4rPgrvmYbBrxZCK_GwXvDVL3
Stanford machine learning theory https://www.youtube.com/playlist?list=PLoROMvodv4rP8nAmISxFINlGKSK4rbLKh
Stanford computer vision https://www.youtube.com/playlist?list=PLkt2uSq6rBVctENoVBg1TpCC7OQi31AlC
https://www.youtube.com/playlist?list=PLSVEhWrZWDHQTBmWZufjxpw3s8sveJtnJ
Stanford statistics https://www.youtube.com/playlist?list=PLoROMvodv4rOpr_A7B9SriE_iZmkanvUg
Stanford methods in AI https://www.youtube.com/playlist?list=PLoROMvodv4rO1NB9TD4iUZ3qghGEGtqNX
https://www.youtube.com/playlist?list=PLrxfgDEc2NxZJcWcrxH3jyjUUrJlnoyzX
Stanford MIT robotics https://www.youtube.com/playlist?list=PLkx8KyIQkMfUmB3j-DyP58ThDXM7enA8x https://www.youtube.com/playlist?list=PLkx8KyIQkMfUSDs2hvTWzaq-cxGl8Ha69 https://www.youtube.com/playlist?list=PL65CC0384A1798ADF https://www.youtube.com/playlist?list=PLoROMvodv4rMeercb-kvGLUrOq4HR6BZD https://www.youtube.com/playlist?list=PLN1iOWWHLJz3ndzRIvpbby75G2_2pYYrL
MIT machine learning https://www.youtube.com/playlist?list=PLxC_ffO4q_rW0bqQB80_vcQB09HOA3ClV https://www.youtube.com/playlist?list=PLnvKubj2-I2LhIibS8TOGC42xsD3-liux
MIT efficient machine learning https://www.youtube.com/playlist?list=PL80kAHvQbh-pT4lCkDT53zT8DKmhE0idB
MIT linear algebra in machine learning https://www.youtube.com/playlist?list=PLUl4u3cNGP63oMNUHXqIUcrkS2PivhN3k
Principles of Deep Learning Theory https://arxiv.org/abs/2106.10165 https://www.youtube.com/watch?v=YzR2gZrsdJc https://www.youtube.com/watch?v=pad023JIXVA
Active Inference book https://mitpress.mit.edu/9780262045353/active-inference/
Geometric deep learning https://geometricdeeplearning.com/
Mechanistic intepretability https://www.neelnanda.io/mechanistic-interpretability
Topological data analysis https://www.youtube.com/playlist?list=PLzERW_Obpmv_UW7RgbZW4Ebhw87BcoXc7

Deep dives

State of the art

  • AI Index Report 2024 – Artificial Intelligence Index Top 10 Takeaways:

  • AI beats humans on some tasks, but not on all. AI has surpassed human performance on several benchmarks, including some in image classification, visual reasoning, and English understanding. Yet it trails behind on more complex tasks like competition-level mathematics, visual commonsense reasoning and planning.

  • Industry continues to dominate frontier AI research. In 2023, industry produced 51 notable machine learning models, while academia contributed only 15. There were also 21 notable models resulting from industry-academia collaborations in 2023, a new high.

  • Frontier models get way more expensive. According to AI Index estimates, the training costs of state-of-the-art AI models have reached unprecedented levels. For example, OpenAI’s GPT-4 used an estimated $78 million worth of compute to train, while Google’s Gemini Ultra cost $191 million for compute.

  • The United States leads China, the EU, and the U.K. as the leading source of top AI models. In 2023, 61 notable AI models originated from U.S.-based institutions, far outpacing the European Union’s 21 and China’s 15.

  • Robust and standardized evaluations for LLM responsibility are seriously lacking. New research from the AI Index reveals a significant lack of standardization in responsible AI reporting. Leading developers, including OpenAI, Google, and Anthropic, primarily test their models against different responsible AI benchmarks. This practice complicates efforts to systematically compare the risks and limitations of top AI models.

  • Generative AI investment skyrockets. Despite a decline in overall AI private investment last year, funding for generative AI surged, nearly octupling from 2022 to reach $25.2 billion. Major players in the generative AI space, including OpenAI, Anthropic, Hugging Face, and Inflection, reported substantial fundraising rounds.

  • The data is in: AI makes workers more productive and leads to higher quality work. In 2023, several studies assessed AI’s impact on labor, suggesting that AI enables workers to complete tasks more quickly and to improve the quality of their output. These studies also demonstrated AI’s potential to bridge the skill gap between low- and high-skilled workers. Still, other studies caution that using AI without proper oversight can lead to diminished performance.

  • Scientific progress accelerates even further, thanks to AI. In 2022, AI began to advance scientific discovery. 2023, however, saw the launch of even more significant science-related AI applications— from AlphaDev, which makes algorithmic sorting more efficient, to GNoME, which facilitates the process of materials discovery.

  • The number of AI regulations in the United States sharply increases. The number of AIrelated regulations in the U.S. has risen significantly in the past year and over the last five years. In 2023, there were 25 AI-related regulations, up from just one in 2016. Last year alone, the total number of AI-related regulations grew by 56.3%.

  • People across the globe are more cognizant of AI’s potential impact—and more nervous. A survey from Ipsos shows that, over the last year, the proportion of those who think AI will dramatically affect their lives in the next three to five years has increased from 60% to 66%. Moreover, 52% express nervousness toward AI products and services, marking a 13 percentage point rise from 2022. In America, Pew data suggests that 52% of Americans report feeling more concerned than excited about AI, rising from 37% in 2022.

Brain storming

And humans may be podobně reduktivně "just doplňovač predikcí input signálů that are compared to actual signals" (using a version of bayesian inference) Predictive coding nebo "jenom bioeletrika a biochemie" nebo "jenom částice"

Teď je asi největší limitace do AI systémů nacpat víc complex systematic coherent reasoning, planning, generalizing, agentnost (autonomita), memory, factual groundedness, online learning, human like etický uvažování, ovládatelnost, což mají celkem weak na složitější tasky když se škálují jejich capabilities, ale děláme v tom progress, buď přes composing LLMs v multiagent systémech, škálování, kvalitnější data a trénování, šťourání jak uvnitř fungujou a tak je ovládat, přes lepší matematický modely jak learning funguje, nebo poupravenou nebo překopanou architekturu, apod.... a nebo se ještě řeší fungující robotika... a všechny top AI laby na těchto věcech v různých mírách pracují/investují. Tady jsou nějaký práce:

Survey of LLMs: [2312.03863] Efficient Large Language Models: A Survey, [2311.10215] Predictive Minds: LLMs As Atypical Active Inference Agents, A Systematic Survey of Prompt Engineering in Large Language Models: Techniques and Applications

Reasoning: Human-like systematic generalization through a meta-learning neural network | Nature, [2305.20050] Let's Verify Step by Step, [2302.00923] Multimodal Chain-of-Thought Reasoning in Language Models, [2310.09158] Learning To Teach Large Language Models Logical Reasoning, [2303.09014] ART: Automatic multi-step reasoning and tool-use for large language models, AlphaGeometry: An Olympiad-level AI system for geometry - Google DeepMind

Robotics: Mobile ALOHA - A Smart Home Robot - Compilation of Autonomous Skills - YouTube, Eureka! Extreme Robot Dexterity with LLMs | NVIDIA Research Paper - YouTube, Shaping the future of advanced robotics - Google DeepMind, Optimus - Gen 2 - YouTube, Atlas Struts - YouTube, Figure Status Update - AI Trained Coffee Demo - YouTube, Curiosity-Driven Learning of Joint Locomotion and Manipulation Tasks - YouTube

Multiagent systems: [2402.01680] Large Language Model based Multi-Agents: A Survey of Progress and Challenges

Pozměněný/alternativní architektury: Mamba (deep learning architecture) - Wikipedia, [2305.13048] RWKV: Reinventing RNNs for the Transformer Era, V-JEPA: The next step toward advanced machine intelligence, Active Inference

Agency: [2305.16291] Voyager: An Open-Ended Embodied Agent with Large Language Models, [2309.07864] The Rise and Potential of Large Language Model Based Agents: A Survey, Agents | Langchain, GitHub - THUDM/AgentBench: A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24), [2401.12917] Active Inference as a Model of Agency, CAN AI THINK ON ITS OWN? - YouTube, Artificial Curiosity Since 1990

Factual groundedness: [2312.10997] Retrieval-Augmented Generation for Large Language Models: A Survey, Perplexity, ChatGPT - Consensus

Memory:
větší context window Gemini 10 mil token context window, nebo vektorový databáze

Online learning: <https://en.wikipedia.org/wiki/Online_machine_learning>

Meta learning: https://en.wikipedia.org/wiki/Meta-learning_(computer_science)

Planning: [2402.01817] LLMs Can't Plan, But Can Help Planning in LLM-Modulo Frameworks, [2401.11708v1] Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs, [2305.16151] Understanding the Capabilities of Large Language Models for Automated Planning

Generalizing: [2402.10891] Instruction Diversity Drives Generalization To Unseen Tasks, Automated discovery of algorithms from data | Nature Computational Science, [2402.09371] Transformers Can Achieve Length Generalization But Not Robustly, [2310.16028] What Algorithms can Transformers Learn? A Study in Length Generalization, [2307.04721] Large Language Models as General Pattern Machines, A Tutorial on Domain Generalization | Proceedings of the Sixteenth ACM International Conference on Web Search and Data Mining, [2311.06545] Understanding Generalization via Set Theory, [2310.08661] Counting and Algorithmic Generalization with Transformers, Neural Networks on the Brink of Universal Prediction with DeepMind’s Cutting-Edge Approach | Synced, [2401.14953] Learning Universal Predictors, Spectral bias and task-model alignment explain generalization in kernel regression and infinitely wide neural networks | Nature Communications

Je dost možný (a velký % researchers si myslí) že research snažící se ovládat tyhle šílený inscrutabe matrices nemá dostatečně rychlej development v porovnávání s capabilities researchem (zvětšování množství věcí čeho ty systémy jsou schopny)

Potom nemáme tušení jak vypnout behaviors s dosavadníma metodama Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training Anthropic, což teď poslední dny šlo vidět s tím jak GPT4 začlo outputovat totální chaos po updatu OpenAI’s ChatGPT Went Completely Off the Rails for Hours, Gemini bylo víc woke než bylo intended, nebo každou chvíl vidím novej jailbreak co obchází zábrany :PepeLaugh~1: [2307.15043] Universal and Transferable Adversarial Attacks on Aligned Language Models.

Co se týče definicí AGI, tohle je dobrý od DeepMindu Levels of AGI: Operationalizing Progress on the Path to AGI, nebo ještě mám rád i když celkem vague tak celkem dobrou definici od OpenAI: Highly autonomous systems that outperform humans at most economically valuable work, nebo tohle je fajn thread různých definicí a jejich výhod a nevýhod 9 definitions of Artificial General Intelligence (AGI) and why they are flawed, nebo ještě Universal Intelligence: A Definition of Machine Intelligence, nebo Karl Friston má dobrý definice KARL FRISTON - INTELLIGENCE 3.0)

Pokud chcete jiný zdroj predikcí o AI od dalších AI lidí než od Metaculusu, LessWrong/EA, Singularity komunity a různých vlivných v top AI labech, co mají hodně short timelines v příštích 10 letech. Singularity Predictions 2024 by some people big in the field, Metaculus: When will the first weakly general AI system be devised, tested, and publicly announced?

Tak existují ještě tyhle priority a predikce, jejíž intervaly se v těchto questionares každým rokem ~dva krát zmenšují:
AI experts make predictions for 2040. I was a little surprised. | Science News, Thousands of AI Authors on the Future of AI:
"In the largest survey of its kind, 2,778 researchers who had published in top-tier artificial intelligence (AI) venues gave predictions on the pace of AI progress and the nature and impacts of advanced AI systems The aggregate forecasts give at least a 50% chance of AI systems achieving several milestones by 2028, including autonomously constructing a payment processing site from scratch, creating a song indistinguishable from a new song by a popular musician, and autonomously downloading and fine-tuning a large language model. If science continues undisrupted, the chance of unaided machines outperforming humans in every possible task was estimated at 10% by 2027, and 50% by 2047. The latter estimate is 13 years earlier than that reached in a similar survey we conducted only one year earlier [Grace et al., 2022]. However, the chance of all human occupations becoming fully automatable was forecast to reach 10% by 2037, and 50% as late as 2116 (compared to 2164 in the 2022 survey).
Most respondents expressed substantial uncertainty about the long-term value of AI progress: While 68.3% thought good outcomes from superhuman AI are more likely than bad, of these net optimists 48% gave at least a 5% chance of extremely bad outcomes such as human extinction, and 59% of net pessimists gave 5% or more to extremely good outcomes. Between 38% and 51% of respondents gave at least a 10% chance to advanced AI leading to outcomes as bad as human extinction. More than half suggested that "substantial" or "extreme" concern is warranted about six different AI-related scenarios, including misinformation, authoritarian control, and inequality. There was disagreement about whether faster or slower AI progress would be better for the future of humanity. However, there was broad agreement that research aimed at minimizing potential risks from AI systems ought to be prioritized more."

Additional resources

AI

  • Artificial Intelligence (AI) is a rapidly evolving field with numerous branches and sub-disciplines. Here's a comprehensive list of various branches of AI:

1. Machine Learning

  • Supervised Learning
  • Unsupervised Learning
  • Reinforcement Learning
  • Deep Learning
  • Neural Networks
  • Decision Trees
  • Support Vector Machines
  • Ensemble Methods
  • Clustering
  • Feature Engineering
  • Dimensionality Reduction
  • Model Selection and Training
  • Transfer Learning
  • Federated Learning

2. Natural Language Processing (NLP)

  • Speech Recognition
  • Text-to-Speech
  • Sentiment Analysis
  • Machine Translation
  • Word Embeddings
  • Named Entity Recognition
  • Part-of-Speech Tagging
  • Language Modeling
  • Text Summarization
  • Dialog Systems and Chatbots
  • Question Answering Systems
  • Natural Language Understanding
  • Natural Language Generation

3. Computer Vision

  • Image Recognition and Classification
  • Object Detection
  • Face Recognition
  • Optical Character Recognition (OCR)
  • Image Segmentation
  • Pattern Recognition
  • Motion Analysis and Tracking
  • Scene Reconstruction
  • Image Enhancement
  • 3D Vision
  • Augmented Reality

4. Robotics

  • Robotic Process Automation (RPA)
  • Humanoid Robots
  • Autonomous Vehicles
  • Drone Robotics
  • Industrial Robotics
  • Swarm Robotics
  • Soft Robotics
  • Rehabilitation Robotics
  • Robotic Surgery
  • Human-Robot Interaction

5. Knowledge Representation and Reasoning

  • Expert Systems
  • Ontologies
  • Semantic Networks
  • Fuzzy Logic Systems
  • Rule-Based Systems
  • Commonsense Reasoning
  • Case-Based Reasoning
  • Qualitative Reasoning
  • Deductive Reasoning

6. Planning and Scheduling

  • Automatic Planning
  • Decision Support Systems
  • Multi-agent Systems
  • Game Theory
  • Constraint Satisfaction
  • Resource Allocation
  • Workflow Management

7. Search and Optimization

  • Genetic Algorithms
  • Evolutionary Computing
  • Swarm Intelligence
  • Simulated Annealing
  • Hill Climbing
  • Pathfinding Algorithms
  • Particle Swarm Optimization

8. Artificial Neural Networks

  • Convolutional Neural Networks (CNN)
  • Recurrent Neural Networks (RNN)
  • Long Short-Term Memory Networks (LSTM)
  • Generative Adversarial Networks (GAN)
  • Deep Belief Networks
  • Autoencoders
  • Radial Basis Function Networks
  • Transformers
  • Linear-Time Sequence Modeling with Selective State Spaces

9. Data Mining and Big Data

  • Predictive Analytics
  • Data Warehousing
  • Big Data Analytics
  • Data Visualization
  • Association Rule Learning
  • Anomaly Detection

10. Affective Computing

  • Emotion Recognition
  • Affective Interfaces
  • Emotional AI
  • Human Affective Response Analysis

11. AI Ethics and Safety

  • Explainable AI
  • Fairness and Bias in AI
  • AI Governance
  • Privacy-Preserving AI
  • AI Safety and Robustness
  • Trustworthy AI

12. Cognitive Computing

  • Cognitive Modeling
  • Human-Centered AI
  • Neuromorphic Computing
  • Cognitive Robotics
  • Hybrid Intelligent Systems

13. AI in Healthcare

  • Medical Image Analysis
  • Predictive Diagnostics
  • Drug Discovery
  • Personalized Medicine
  • Patient Data Analysis

14. AI in Business

  • Customer Relationship Management
  • Business Intelligence
  • Market Analysis
  • Supply Chain Optimization
  • AI in Finance and Trading

15. AI in Education

  • Adaptive Learning Systems
  • Educational Data Mining
  • AI Tutors
  • Learning Analytics
  • Curriculum Design

16. Quantum AI

  • Quantum Machine Learning
  • Quantum Computing for AI
  • Quantum Optimization

I want to learn and have a map of as much of the mathematical theory and practice of as much methods as possible used everywhere for example in:
- statistical methods (frequentist, bayesian statistics,...)
- machine learning (supervised learning (classification, all sorts of regression), unsupervised learning (clusering, dimensionality reduction,...), semisupervised learning, reinforcement learning, ensemble methods,...)
- deep learning (all variations and combinations of classic neural nets, convolutional NNs, recurrent NNs, LSTMs, GANs, selforganizing maps, deep belief networks, deep RL, graph NNs, neural turing machines, all variations of transformers, rwkw, xLSTM, diffusion,...)
- symbolic methods, neurosymbolics, statespace models, graph analysis, other stuff in natural language processing, computer vision, signal processing, anomaly detection, recommender systems, different optimization algorithms and metaheuristics, metalearning,...
etc. etc. etc.
All of it includes essentially infinite amount of infinite rabbitholes, but it's worth it

Map of algorithms for extracting patterns from data

  1. Statistical Methods
  2. Descriptive Statistics
  3. Central Tendency (Mean, Median, Mode, Geometric Mean, Harmonic Mean)
  4. Dispersion (Range, Variance, Standard Deviation, Coefficient of Variation, Quartiles, Interquartile Range)
  5. Skewness and Kurtosis
  6. Inferential Statistics
  7. Hypothesis Testing (Z-test, t-test, F-test, Chi-Square Test, ANOVA, MANOVA, ANCOVA)
  8. Confidence Intervals
  9. Non-parametric Tests (Mann-Whitney U, Wilcoxon Signed-Rank, Kruskal-Wallis, Friedman)
  10. Regression Analysis
  11. Linear Regression (Simple, Multiple)
  12. Logistic Regression (Binary, Multinomial, Ordinal)
  13. Polynomial Regression
  14. Stepwise Regression
  15. Ridge Regression
  16. Lasso Regression
  17. Elastic Net Regression
  18. Bayesian Statistics
  19. Bayesian Inference
  20. Naive Bayes Classifier
  21. Bayesian Networks
  22. Markov Chain Monte Carlo (MCMC) Methods
  23. Survival Analysis
  24. Kaplan-Meier Estimator
  25. Cox Proportional Hazards Model
  26. Spatial Statistics
  27. Kriging
  28. Spatial Autocorrelation (Moran's I, Geary's C)

  29. Machine Learning

  30. Supervised Learning
  31. Classification
  32. Decision Trees & Random Forests
  33. Naive Bayes (Gaussian, Multinomial, Bernoulli)
  34. Support Vector Machines (SVM) (Linear, RBF, Polynomial)
  35. k-Nearest Neighbors (k-NN)
  36. Logistic Regression
  37. Neural Networks (Feedforward, Convolutional, Recurrent)
  38. Gradient Boosting Machines (GBM)
  39. AdaBoost
  40. XGBoost
  41. LightGBM
  42. CatBoost
  43. Regression
  44. Linear Regression
  45. Polynomial Regression
  46. Support Vector Regression (SVR)
  47. Decision Trees & Random Forests
  48. Neural Networks (Feedforward, Convolutional, Recurrent)
  49. Gradient Boosting Machines (GBM)
  50. AdaBoost
  51. XGBoost
  52. LightGBM
  53. CatBoost
  54. Unsupervised Learning
  55. Clustering
  56. k-Means
  57. Mini-Batch k-Means
  58. Hierarchical Clustering (Agglomerative, Divisive)
  59. DBSCAN
  60. OPTICS
  61. Mean Shift
  62. Gaussian Mixture Models
  63. Fuzzy C-Means
  64. Dimensionality Reduction
  65. Principal Component Analysis (PCA)
  66. Kernel PCA
  67. Incremental PCA
  68. t-SNE
  69. UMAP
  70. Isomap
  71. Locally Linear Embedding (LLE)
  72. Independent Component Analysis (ICA)
  73. Non-Negative Matrix Factorization (NMF)
  74. Latent Dirichlet Allocation (LDA)
  75. Autoencoders (Vanilla, Variational, Denoising)
  76. Association Rule Mining
  77. Apriori
  78. FP-Growth
  79. ECLAT
  80. Semi-Supervised Learning
  81. Self-Training
  82. Co-Training
  83. Graph-Based Methods
  84. Transductive SVM
  85. Generative Models
  86. Reinforcement Learning
  87. Q-Learning
  88. SARSA
  89. Deep Q Networks (DQN)
  90. Policy Gradients (REINFORCE, Actor-Critic)
  91. Proximal Policy Optimization (PPO)
  92. Monte Carlo Methods
  93. Temporal Difference Learning
  94. AlphaZero
  95. Ensemble Methods
  96. Bagging
  97. Boosting (AdaBoost, Gradient Boosting, XGBoost, LightGBM, CatBoost)
  98. Stacking
  99. Voting (Majority, Weighted, Soft)
  100. Random Subspace Method
  101. Rotation Forests

  102. Deep Learning

  103. Feedforward Neural Networks
  104. Convolutional Neural Networks (CNN)
  105. LeNet
  106. AlexNet
  107. VGGNet
  108. ResNet
  109. Inception
  110. DenseNet
  111. EfficientNet
  112. Recurrent Neural Networks (RNN)
  113. Long Short-Term Memory (LSTM)
  114. Gated Recurrent Units (GRU)
  115. Bidirectional RNNs
  116. Transformers
  117. Attention Mechanism
  118. Self-Attention
  119. Multi-Head Attention
  120. BERT
  121. GPT
  122. Transformer-XL
  123. XLNet
  124. Autoencoders
  125. Vanilla Autoencoders
  126. Variational Autoencoders (VAE)
  127. Denoising Autoencoders
  128. Sparse Autoencoders
  129. Generative Adversarial Networks (GANs)
  130. Vanilla GANs
  131. Deep Convolutional GANs (DCGANs)
  132. Conditional GANs
  133. Wasserstein GANs (WGANs)
  134. Cycle GANs
  135. StyleGANs
  136. Self-Organizing Maps (SOMs)
  137. Deep Belief Networks (DBNs)
  138. Deep Reinforcement Learning
  139. Deep Q Networks (DQN)
  140. Double DQN
  141. Dueling DQN
  142. Deep Deterministic Policy Gradient (DDPG)
  143. Asynchronous Advantage Actor-Critic (A3C)

  144. Time Series Analysis

  145. Exploratory Data Analysis
  146. Seasonality
  147. Trend
  148. Cyclicality
  149. Autocorrelation
  150. Partial Autocorrelation
  151. Smoothing Techniques
  152. Moving Averages (Simple, Weighted, Exponential)
  153. Holt-Winters (Additive, Multiplicative)
  154. Kalman Filter
  155. Decomposition Methods
  156. Classical Decomposition (Additive, Multiplicative)
  157. STL Decomposition
  158. Regression-based Methods
  159. Linear Regression
  160. Autoregressive Models (AR)
  161. Moving Average Models (MA)
  162. Autoregressive Moving Average Models (ARMA)
  163. Autoregressive Integrated Moving Average Models (ARIMA)
  164. Seasonal ARIMA (SARIMA)
  165. Vector Autoregression (VAR)
  166. State Space Models
  167. Exponential Smoothing State Space Models (ETS)
  168. Structural Time Series Models
  169. Dynamic Linear Models (DLMs)
  170. Machine Learning Methods
  171. Prophet
  172. Recurrent Neural Networks (RNNs)
  173. Long Short-Term Memory (LSTM)
  174. Gated Recurrent Units (GRUs)
  175. Temporal Convolutional Networks (TCNs)
  176. XGBoost
  177. Ensemble Methods
  178. Bagging
  179. Boosting
  180. Stacking
  181. Anomaly Detection
  182. Statistical Process Control
  183. Isolation Forests
  184. Robust PCA
  185. Causality Analysis
  186. Granger Causality
  187. Vector Autoregression (VAR)
  188. Convergent Cross Mapping (CCM)

  189. Anomaly Detection

  190. Statistical Methods
  191. Z-Score
  192. Interquartile Range (IQR)
  193. Mahalanobis Distance
  194. Kernel Density Estimation (KDE)
  195. Clustering-Based Methods
  196. k-Means
  197. DBSCAN
  198. Density-Based Methods
  199. Local Outlier Factor (LOF)
  200. Connectivity-Based Outlier Factor (COF)
  201. Subspace Outlier Detection (SOD)
  202. Distance-Based Methods
  203. k-Nearest Neighbors (k-NN)
  204. Ensemble Methods
  205. Isolation Forest
  206. Feature Bagging
  207. Subsampling
  208. One-Class Classification
  209. One-Class SVM
  210. Support Vector Data Description (SVDD)
  211. Autoencoder-based Methods
  212. Probabilistic Methods
  213. Gaussian Mixture Models (GMMs)
  214. Hidden Markov Models (HMMs)
  215. Bayesian Networks

  216. Natural Language Processing (NLP)

  217. Text Preprocessing
  218. Tokenization
  219. Stop Word Removal
  220. Stemming & Lemmatization
  221. Part-of-Speech (POS) Tagging
  222. Named Entity Recognition (NER)
  223. Parsing
  224. Text Representation
  225. Bag-of-Words (BoW)
  226. TF-IDF
  227. Word Embeddings (Word2Vec, GloVe, FastText)
  228. Sentence Embeddings (Doc2Vec, Sent2Vec)
  229. Contextual Embeddings (ELMo, BERT, GPT)
  230. Text Classification
  231. Naive Bayes
  232. Support Vector Machines (SVM)
  233. Logistic Regression
  234. Decision Trees & Random Forests
  235. Neural Networks (CNNs, RNNs, Transformers)
  236. Sequence Labeling
  237. Hidden Markov Models (HMMs)
  238. Conditional Random Fields (CRFs)
  239. Recurrent Neural Networks (RNNs)
  240. Transformers
  241. Topic Modeling
  242. Latent Dirichlet Allocation (LDA)
  243. Non-Negative Matrix Factorization (NMF)
  244. Latent Semantic Analysis (LSA)
  245. Hierarchical Dirichlet Process (HDP)
  246. Text Summarization
  247. Extractive Methods (TextRank, LexRank)
  248. Abstractive Methods (Seq2Seq Models, Transformers)
  249. Machine Translation
  250. Statistical Machine Translation (SMT)
  251. Neural Machine Translation (NMT)
  252. Seq2Seq Models
  253. Attention Mechanisms
  254. Transformers
  255. Sentiment Analysis
  256. Lexicon-based Methods
  257. Machine Learning Methods (Naive Bayes, SVM, Logistic Regression)
  258. Deep Learning Methods (CNNs, RNNs, Transformers)
  259. Language Modeling
  260. N-gram Models
  261. Neural Language Models (RNNs, LSTMs, GRUs)
  262. Transformers (GPT, BERT)
  263. Text Generation
  264. Rule-based Methods
  265. Statistical Language Models
  266. Neural Language Models (RNNs, LSTMs, GRUs)
  267. Transformers (GPT, BERT)
  268. Information Retrieval
  269. Boolean Models
  270. Vector Space Models (TF-IDF)
  271. Probabilistic Models (BM25)
  272. Learning to Rank (LTR)
  273. Named Entity Recognition (NER)
  274. Rule-based Methods
  275. Machine Learning Methods (CRFs, HMMs)
  276. Deep Learning Methods (BiLSTM-CRF, Transformers)
  277. Relationship Extraction
  278. Pattern-based Methods
  279. Machine Learning Methods (SVMs, CRFs)
  280. Deep Learning Methods (CNNs, RNNs, Transformers)
  281. Coreference Resolution
  282. Rule-based Methods
  283. Machine Learning Methods (Mention-Pair, Entity-Mention)
  284. Deep Learning Methods (Mention Ranking, End-to-End Models)

  285. Computer Vision

  286. Image Preprocessing
  287. Pixel-level Operations (Scaling, Cropping, Rotation, Flipping)
  288. Filtering (Gaussian, Median, Bilateral)
  289. Edge Detection (Sobel, Canny, Laplacian)
  290. Morphological Operations (Erosion, Dilation, Opening, Closing)
  291. Feature Extraction
  292. Scale-Invariant Feature Transform (SIFT)
  293. Speeded Up Robust Features (SURF)
  294. Oriented FAST and Rotated BRIEF (ORB)
  295. Histogram of Oriented Gradients (HOG)
  296. Local Binary Patterns (LBP)
  297. Object Detection
  298. Viola-Jones
  299. Sliding Window
  300. Deformable Part Models (DPM)
  301. Region-based CNN (R-CNN, Fast R-CNN, Faster R-CNN)
  302. You Only Look Once (YOLO)
  303. Single Shot MultiBox Detector (SSD)
  304. RetinaNet
  305. Semantic Segmentation
  306. Fully Convolutional Networks (FCNs)
  307. U-Net
  308. DeepLab
  309. Mask R-CNN
  310. Instance Segmentation
  311. Mask R-CNN
  312. PANet
  313. Image Classification
  314. Convolutional Neural Networks (CNNs)
  315. Transfer Learning (VGG, ResNet, Inception, DenseNet, EfficientNet)
  316. Ensemble Methods (Bagging, Boosting)
  317. Object Tracking
  318. Kalman Filter
  319. Particle Filter
  320. Optical Flow
  321. Siamese Networks
  322. Correlation Filter
  323. Pose Estimation
  324. Deformable Part Models (DPM)
  325. Convolutional Pose Machines (CPMs)
  326. Stacked Hourglass Networks
  327. OpenPose
  328. Face Recognition
  329. Eigenfaces
  330. Local Binary Patterns Histograms (LBPH)
  331. FaceNet
  332. DeepFace
  333. DeepID
  334. Generative Models
  335. Variational Autoencoders (VAEs)
  336. Generative Adversarial Networks (GANs)
  337. Neural Style Transfer
  338. Deep Dream
  339. 3D Computer Vision
  340. Structure from Motion (SfM)
  341. Simultaneous Localization and Mapping (SLAM)
  342. Stereo Vision
  343. Point Cloud Processing
  344. Voxel-based Methods

  345. Graph Analytics

  346. Graph Representation
  347. Adjacency Matrix
  348. Adjacency List
  349. Edge List
  350. Incidence Matrix
  351. Graph Traversal
  352. Breadth-First Search (BFS)
  353. Depth-First Search (DFS)
  354. Shortest Path Algorithms
  355. Dijkstra's Algorithm
  356. Bellman-Ford Algorithm
  357. A* Search
  358. Floyd-Warshall Algorithm
  359. Centrality Measures
  360. Degree Centrality
  361. Betweenness Centrality
  362. Closeness Centrality
  363. Eigenvector Centrality
  364. PageRank
  365. HITS (Hubs and Authorities)
  366. Community Detection
  367. Girvan-Newman Algorithm
  368. Louvain Algorithm
  369. Infomap
  370. Spectral Clustering
  371. Stochastic Block Models
  372. Link Prediction
  373. Common Neighbors
  374. Jaccard Coefficient
  375. Adamic-Adar Index
  376. Preferential Attachment
  377. Katz Index
  378. Matrix Factorization
  379. Graph Embeddings
  380. DeepWalk
  381. node2vec
  382. Graph Convolutional Networks (GCNs)
  383. GraphSAGE
  384. Graph Attention Networks (GATs)
  385. Subgraph Matching
  386. Ullmann's Algorithm
  387. VF2 Algorithm
  388. Graph Kernels
  389. Network Motifs
  390. Motif Counting
  391. Motif Discovery
  392. Temporal Graph Analysis
  393. Temporal Motifs
  394. Dynamic Community Detection
  395. Temporal Link Prediction
  396. Graph Neural Networks (GNNs)
  397. Graph Convolutional Networks (GCNs)
  398. Graph Attention Networks (GATs)
  399. Graph Recurrent Networks (GRNs)
  400. Graph Autoencoders
  401. Graph Generative Models

  402. Recommender Systems

  403. Content-based Filtering
  404. TF-IDF
  405. Cosine Similarity
  406. Jaccard Similarity
  407. Collaborative Filtering
  408. User-based Collaborative Filtering
  409. Item-based Collaborative Filtering
  410. Matrix Factorization (Singular Value Decomposition, Non-Negative Matrix Factorization)
  411. Factorization Machines
  412. Probabilistic Matrix Factorization
  413. Hybrid Methods
  414. Weighted Hybrid
  415. Switching Hybrid
  416. Cascade Hybrid
  417. Feature Combination
  418. Meta-level
  419. Context-Aware Recommender Systems
  420. Contextual Pre-filtering
  421. Contextual Post-filtering
  422. Contextual Modeling
  423. Deep Learning-based Recommender Systems
  424. Neural Collaborative Filtering
  425. Deep Matrix Factorization
  426. Autoencoders
  427. Convolutional Neural Networks (CNNs)
  428. Recurrent Neural Networks (RNNs)
  429. Graph Neural Networks (GNNs)
  430. Evaluation Metrics
  431. Precision and Recall
  432. Mean Average Precision (MAP)
  433. Normalized Discounted Cumulative Gain (NDCG)
  434. Mean Reciprocal Rank (MRR)
  435. Coverage
  436. Diversity
  437. Novelty
  438. Serendipity

  439. Optimization Algorithms

  440. Gradient Descent
  441. Batch Gradient Descent
  442. Stochastic Gradient Descent (SGD)
  443. Mini-batch Gradient Descent
  444. Newton's Method
  445. Quasi-Newton Methods
  446. BFGS
  447. L-BFGS
  448. Conjugate Gradient Methods
  449. Momentum
  450. Nesterov Accelerated Gradient (NAG)
  451. Adagrad
  452. Adadelta
  453. RMSprop
  454. Adam
  455. AdaMax
  456. Nadam
  457. AMSGrad
  458. Evolutionary Algorithms
  459. Genetic Algorithms
  460. Evolutionary Strategies
  461. Particle Swarm Optimization (PSO)
  462. Ant Colony Optimization (ACO)
  463. Differential Evolution
  464. Swarm Intelligence Algorithms
  465. Artificial Bee Colony (ABC)
  466. Firefly Algorithm
  467. Cuckoo Search
  468. Bat Algorithm
  469. Simulated Annealing
  470. Tabu Search
  471. Hill Climbing
  472. Gradient-Free Optimization
  473. Nelder-Mead Method
  474. Pattern Search
  475. Bayesian Optimization
  476. Constrained Optimization
  477. Lagrange Multipliers
  478. Karush-Kuhn-Tucker (KKT) Conditions
  479. Interior Point Methods
  480. Penalty Methods
  481. Multi-Objective Optimization
  482. Weighted Sum Method
  483. ε-Constraint Method
  484. Pareto Optimization
  485. Non-dominated Sorting Genetic Algorithm (NSGA-II)
  486. Strength Pareto Evolutionary Algorithm (SPEA2)

This comprehensive map covers a wide range of algorithms and techniques used for extracting patterns and insights from various types of data, including tabular data, time series data, text data, image data, and graph data. It encompasses statistical methods, machine learning algorithms (both traditional and deep learning-based), natural language processing techniques, computer vision algorithms, graph analytics, recommender systems, and optimization algorithms.

The choice of algorithm depends on the specific problem at hand, the nature and structure of the data, the desired outcome, and the trade-offs between accuracy, interpretability, scalability, and computational efficiency. It is essential to have a good understanding of the strengths and limitations of each algorithm and to experiment with different approaches to find the most suitable one for a given task.

Furthermore, data preprocessing, feature engineering, model selection, hyperparameter tuning, and model evaluation are crucial steps in the data analysis pipeline that can significantly impact the performance of the chosen algorithm. It is also important to consider the ethical implications and potential biases associated with the use of these algorithms, especially in sensitive domains such as healthcare, finance, and criminal justice.

Additional metadata

  • processed #processing #toprocess #important #short #long #casual #focus

  • Unfinished: #metadata #tags