Remove content from AI engineer roadmap

pull/7327/head
Kamran Ahmed 2 weeks ago
parent 6461ccaf59
commit be221a2b14
  1. 8
      src/data/roadmaps/ai-engineer/content/adding-end-user-ids-in-prompts@4Q5x2VCXedAWISBXUIyin.md
  2. 9
      src/data/roadmaps/ai-engineer/content/agents-usecases@778HsQzTuJ_3c9OSn5DmH.md
  3. 9
      src/data/roadmaps/ai-engineer/content/ai-agents@9XCxilAQ7FRet7lHQr1gE.md
  4. 9
      src/data/roadmaps/ai-engineer/content/ai-agents@AeHkNU-uJ_gBdo5-xdpEu.md
  5. 9
      src/data/roadmaps/ai-engineer/content/ai-code-editors@XcKeQfpTA5ITgdX51I4y-.md
  6. 4
      src/data/roadmaps/ai-engineer/content/ai-engineer-vs-ml-engineer@jSZ1LhPdhlkW-9QJhIvFs.md
  7. 4
      src/data/roadmaps/ai-engineer/content/ai-safety-and-ethics@8ndKHDJgL_gYwaXC7XMer.md
  8. 4
      src/data/roadmaps/ai-engineer/content/ai-vs-agi@5QdihE1lLpMc3DFrGy46M.md
  9. 4
      src/data/roadmaps/ai-engineer/content/anomaly-detection@AglWJ7gb9rTT2rMkstxtk.md
  10. 8
      src/data/roadmaps/ai-engineer/content/anthropics-claude@hy6EyKiNxk1x84J63dhez.md
  11. 4
      src/data/roadmaps/ai-engineer/content/audio-processing@mxQYB820447DC6kogyZIL.md
  12. 8
      src/data/roadmaps/ai-engineer/content/aws-sagemaker@OkYO-aSPiuVYuLXHswBCn.md
  13. 8
      src/data/roadmaps/ai-engineer/content/azure-ai@3PQVZbcr4neNMRr6CuNzS.md
  14. 8
      src/data/roadmaps/ai-engineer/content/benefits-of-pre-trained-models@1Ga6DbOPc6Crz7ilsZMYy.md
  15. 4
      src/data/roadmaps/ai-engineer/content/bias-and-fareness@lhIU0ulpvDAn1Xc3ooYz_.md
  16. 8
      src/data/roadmaps/ai-engineer/content/capabilities--context-length@vvpYkmycH0_W030E-L12f.md
  17. 8
      src/data/roadmaps/ai-engineer/content/chat-completions-api@_bPTciEA1GT1JwfXim19z.md
  18. 8
      src/data/roadmaps/ai-engineer/content/chroma@dSd2C9lNl-ymmCRT9_ZC3.md
  19. 4
      src/data/roadmaps/ai-engineer/content/chunking@mX987wiZF7p3V_gExrPeX.md
  20. 11
      src/data/roadmaps/ai-engineer/content/code-completion-tools@TifVhqFm1zXNssA8QR3SM.md
  21. 8
      src/data/roadmaps/ai-engineer/content/cohere@a7qsvoauFe5u953I699ps.md
  22. 4
      src/data/roadmaps/ai-engineer/content/conducting-adversarial-testing@Pt-AJmSJrOxKvolb5_HEv.md
  23. 4
      src/data/roadmaps/ai-engineer/content/constraining-outputs-and-inputs@ONLDyczNacGVZGojYyJrU.md
  24. 4
      src/data/roadmaps/ai-engineer/content/cut-off-dates--knowledge@LbB2PeytxRSuU07Bk0KlJ.md
  25. 4
      src/data/roadmaps/ai-engineer/content/dall-e-api@LKFwwjtcawJ4Z12X102Cb.md
  26. 4
      src/data/roadmaps/ai-engineer/content/data-classification@06Xta-OqSci05nV2QMFdF.md
  27. 4
      src/data/roadmaps/ai-engineer/content/development-tools@NYge7PNtfI-y6QWefXJ4d.md
  28. 4
      src/data/roadmaps/ai-engineer/content/embedding@grTcbzT7jKk_sIUwOTZTD.md
  29. 4
      src/data/roadmaps/ai-engineer/content/embeddings@XyEp6jnBSpCxMGwALnYfT.md
  30. 4
      src/data/roadmaps/ai-engineer/content/faiss@JurLbOO1Z8r6C3yUqRNwf.md
  31. 8
      src/data/roadmaps/ai-engineer/content/fine-tuning@15XOFdVp0IC-kLYPXUJWh.md
  32. 4
      src/data/roadmaps/ai-engineer/content/generation@2jJnS9vRYhaS69d6OxrMh.md
  33. 4
      src/data/roadmaps/ai-engineer/content/googles-gemini@oe8E6ZIQWuYvHVbYJHUc1.md
  34. 8
      src/data/roadmaps/ai-engineer/content/hugging-face-hub@YLOdOvLXa5Fa7_mmuvKEi.md
  35. 8
      src/data/roadmaps/ai-engineer/content/hugging-face-models@8XjkRqHOdyH-DbXHYiBEt.md
  36. 8
      src/data/roadmaps/ai-engineer/content/hugging-face-models@EIDbwbdolR_qsNKVDla6V.md
  37. 8
      src/data/roadmaps/ai-engineer/content/hugging-face-tasks@YKIPOiSj_FNtg0h8uaSMq.md
  38. 8
      src/data/roadmaps/ai-engineer/content/hugging-face@v99C5Bml2a6148LCJ9gy9.md
  39. 4
      src/data/roadmaps/ai-engineer/content/image-generation@49BWxYVFpIgZCCqsikH7l.md
  40. 4
      src/data/roadmaps/ai-engineer/content/image-understanding@fzVq4hGoa2gdbIzoyY1Zp.md
  41. 4
      src/data/roadmaps/ai-engineer/content/impact-on-product-development@qJVgKe9uBvXc-YPfvX_Y7.md
  42. 4
      src/data/roadmaps/ai-engineer/content/indexing-embeddings@5TQnO9B4_LTHwqjI7iHB1.md
  43. 9
      src/data/roadmaps/ai-engineer/content/inference-sdk@3kRTzlLNBnXdTsAEXVu_M.md
  44. 4
      src/data/roadmaps/ai-engineer/content/inference@KWjD4xEPhOOYS51dvRLd2.md
  45. 4
      src/data/roadmaps/ai-engineer/content/introduction@_hYN0gEi9BL24nptEtXWU.md
  46. 4
      src/data/roadmaps/ai-engineer/content/know-your-customers--usecases@t1SObMWkDZ1cKqNNlcd9L.md
  47. 4
      src/data/roadmaps/ai-engineer/content/lancedb@rjaCNT3Li45kwu2gXckke.md
  48. 4
      src/data/roadmaps/ai-engineer/content/langchain-for-multimodal-apps@j9zD3pHysB1CBhLfLjhpD.md
  49. 4
      src/data/roadmaps/ai-engineer/content/langchain@ebXXEhNRROjbbof-Gym4p.md
  50. 4
      src/data/roadmaps/ai-engineer/content/limitations-and-considerations@MXqbQGhNM3xpXlMC2ib_6.md
  51. 8
      src/data/roadmaps/ai-engineer/content/llama-index@d0ontCII8KI8wfP-8Y45R.md
  52. 8
      src/data/roadmaps/ai-engineer/content/llamaindex-for-multimodal-apps@akQTCKuPRRelj2GORqvsh.md
  53. 4
      src/data/roadmaps/ai-engineer/content/llms@wf2BSyUekr1S1q6l8kyq6.md
  54. 4
      src/data/roadmaps/ai-engineer/content/manual-implementation@6xaRB34_g0HGt-y1dGYXR.md
  55. 10
      src/data/roadmaps/ai-engineer/content/maximum-tokens@qzvp6YxWDiGakA2mtspfh.md
  56. 4
      src/data/roadmaps/ai-engineer/content/mistral-ai@n-Ud2dXkqIzK37jlKItN4.md
  57. 8
      src/data/roadmaps/ai-engineer/content/models-on-hugging-face@dLEg4IA3F5jgc44Bst9if.md
  58. 8
      src/data/roadmaps/ai-engineer/content/mongodb-atlas@j6bkm0VUgLkHdMDDJFiMC.md
  59. 4
      src/data/roadmaps/ai-engineer/content/multimodal-ai-usecases@sGR9qcro68KrzM8qWxcH8.md
  60. 4
      src/data/roadmaps/ai-engineer/content/multimodal-ai@W7cKPt_UxcUgwp8J6hS4p.md
  61. 8
      src/data/roadmaps/ai-engineer/content/ollama-models@ro3vY_sp6xMQ-hfzO-rc1.md
  62. 8
      src/data/roadmaps/ai-engineer/content/ollama-sdk@TsG_I7FL-cOCSw8gvZH3r.md
  63. 8
      src/data/roadmaps/ai-engineer/content/ollama@rTT2UnvqFO3GH6ThPLEjO.md
  64. 4
      src/data/roadmaps/ai-engineer/content/open-ai-assistant-api@eOqCBgBTKM8CmY3nsWjre.md
  65. 8
      src/data/roadmaps/ai-engineer/content/open-ai-embedding-models@y0qD5Kb4Pf-ymIwW-tvhX.md
  66. 4
      src/data/roadmaps/ai-engineer/content/open-ai-embeddings-api@l6priWeJhbdUD5tJ7uHyG.md
  67. 4
      src/data/roadmaps/ai-engineer/content/open-ai-models@2WbVpRLqwi3Oeqk1JPui4.md
  68. 4
      src/data/roadmaps/ai-engineer/content/open-ai-playground@nyBgEHvUhwF-NANMwkRJW.md
  69. 4
      src/data/roadmaps/ai-engineer/content/open-source-embeddings@apVYIV4EyejPft25oAvdI.md
  70. 4
      src/data/roadmaps/ai-engineer/content/open-vs-closed-source-models@RBwGsq9DngUsl8PrrCbqx.md
  71. 4
      src/data/roadmaps/ai-engineer/content/openai-api@zdeuA4GbdBl2DwKgiOA4G.md
  72. 4
      src/data/roadmaps/ai-engineer/content/openai-assistant-api@mbp2NoL-VZ5hZIIblNBXt.md
  73. 4
      src/data/roadmaps/ai-engineer/content/openai-functions--tools@Sm0Ne5Nx72hcZCdAcC0C2.md
  74. 4
      src/data/roadmaps/ai-engineer/content/openai-models@5ShWZl1QUqPwO-NRGN85V.md
  75. 4
      src/data/roadmaps/ai-engineer/content/openai-moderation-api@ljZLa3yjQpegiZWwtnn_q.md
  76. 4
      src/data/roadmaps/ai-engineer/content/openai-vision-api@CRrqa-dBw1LlOwVbrZhjK.md
  77. 4
      src/data/roadmaps/ai-engineer/content/opensource-ai@a_3SabylVqzzOyw3tZN5f.md
  78. 4
      src/data/roadmaps/ai-engineer/content/performing-similarity-search@ZcbRPtgaptqKqWBgRrEBU.md
  79. 4
      src/data/roadmaps/ai-engineer/content/pinecone@_Cf7S1DCvX7p1_3-tP3C3.md
  80. 4
      src/data/roadmaps/ai-engineer/content/popular-open-source-models@97eu-XxYUH9pYbD_KjAtA.md
  81. 4
      src/data/roadmaps/ai-engineer/content/pre-trained-models@d7fzv_ft12EopsQdmEsel.md
  82. 4
      src/data/roadmaps/ai-engineer/content/pricing-considerations@4GArjDYipit4SLqKZAWDf.md
  83. 4
      src/data/roadmaps/ai-engineer/content/pricing-considerations@DZPM9zjCbYYWBPLmQImxQ.md
  84. 4
      src/data/roadmaps/ai-engineer/content/prompt-engineering@Dc15ayFlzqMF24RqIF_-X.md
  85. 4
      src/data/roadmaps/ai-engineer/content/prompt-injection-attacks@cUyLT6ctYQ1pgmodCKREq.md
  86. 4
      src/data/roadmaps/ai-engineer/content/purpose-and-functionality@WcjX6p-V-Rdd77EL8Ega9.md
  87. 4
      src/data/roadmaps/ai-engineer/content/qdrant@DwOAL5mOBgBiw-EQpAzQl.md
  88. 4
      src/data/roadmaps/ai-engineer/content/rag--implementation@lVhWhZGR558O-ljHobxIi.md
  89. 4
      src/data/roadmaps/ai-engineer/content/rag-usecases@GCn4LGNEtPI0NWYAZCRE-.md
  90. 4
      src/data/roadmaps/ai-engineer/content/rag-vs-fine-tuning@qlBEXrbV88e_wAGRwO9hW.md
  91. 4
      src/data/roadmaps/ai-engineer/content/rag@9JwWIK0Z2MK8-6EQQJsCO.md
  92. 4
      src/data/roadmaps/ai-engineer/content/react-prompting@voDKcKvXtyLzeZdx2g3Qn.md
  93. 4
      src/data/roadmaps/ai-engineer/content/recommendation-systems@HQe9GKy3p0kTUPxojIfSF.md
  94. 4
      src/data/roadmaps/ai-engineer/content/replicate@c0RPhpD00VIUgF4HJgN2T.md
  95. 4
      src/data/roadmaps/ai-engineer/content/retrieval-process@OCGCzHQM2LQyUWmiqe6E0.md
  96. 4
      src/data/roadmaps/ai-engineer/content/robust-prompt-engineering@qmx6OHqx4_0JXVIv8dASp.md
  97. 4
      src/data/roadmaps/ai-engineer/content/roles-and-responsiblities@K9EiuFgPBFgeRxY4wxAmb.md
  98. 4
      src/data/roadmaps/ai-engineer/content/security-and-privacy-concerns@sWBT-j2cRuFqRFYtV_5TK.md
  99. 4
      src/data/roadmaps/ai-engineer/content/semantic-search@eMfcyBxnMY_l_5-8eg6sD.md
  100. 4
      src/data/roadmaps/ai-engineer/content/sentence-transformers@ZV_V6sqOnRodgaw4mzokC.md
  101. Some files were not shown because too many files have changed in this diff Show More

@ -1,7 +1 @@
# Adding end-user IDs in prompts
Sending end-user IDs in your requests can be a useful tool to help OpenAI monitor and detect abuse. This allows OpenAI to provide you with more actionable feedback in the event that they may detect any policy violations in applications.
Visit the following resources to learn more:
- [@official@OpenAI Documentation](https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids)
# Adding end-user IDs in prompts

@ -1,8 +1 @@
# Agents Usecases
AI Agents allow you to automate complex workflows that involve multiple steps and decisions.
Visit the following resources to learn more:
- [@article@What are AI Agents?](https://aws.amazon.com/what-is/ai-agents/)
- [@video@What are AI Agents?](https://www.youtube.com/watch?v=F8NKVhkZZWI)
# Agents Usecases

@ -1,8 +1 @@
# AI Agents
AI Agents are a type of LLM that can be used to automate complex workflows that involve multiple steps and decisions.
Visit the following resources to learn more:
- [@article@What are AI Agents?](https://aws.amazon.com/what-is/ai-agents/)
- [@video@What are AI Agents?](https://www.youtube.com/watch?v=F8NKVhkZZWI)
# AI Agents

@ -1,8 +1 @@
# AI Agents
AI Agents are a type of LLM that can be used to automate complex workflows that involve multiple steps and decisions.
Visit the following resources to learn more:
- [@article@What are AI Agents?](https://aws.amazon.com/what-is/ai-agents/)
- [@video@What are AI Agents?](https://www.youtube.com/watch?v=F8NKVhkZZWI)
# AI Agents

@ -1,8 +1 @@
# AI Code Editors
AI code editors have the first-class support for AI in the editor. You can use AI to generate code, fix bugs, chat with your code, and more.
Visit the following resources to learn more:
- [@website@Cursor](https://cursor.com/)
- [@website@Zed AI](https://zed.dev/ai)
# AI Code Editors

@ -1,3 +1 @@
# AI Engineer vs ML Engineer
AI Engineer differs from an AI Researcher or ML Engineer. AI Engineers focus on leveraging pre-trained models and existing AI technologies to enhance user experiences without the need to train models from scratch.
# AI Engineer vs ML Engineer

@ -1,3 +1 @@
# AI Safety and Ethics
Learn about the principles and guidelines for building safe and ethical AI systems.
# AI Safety and Ethics

@ -1,3 +1 @@
# AI vs AGI
AI (Artificial Intelligence) refers to systems designed to perform specific tasks, like image recognition or language translation, often excelling in those narrow areas. In contrast, AGI (Artificial General Intelligence) would be a system capable of understanding, learning, and applying intelligence across a wide range of tasks, much like a human, and could adapt to new situations without specific programming.
# AI vs AGI

@ -1,3 +1 @@
# Anomaly Detection
Embeddings transform complex data (like text or behavior) into numerical vectors, capturing relationships between data points. These vectors are stored in a vector database, which allows for efficient similarity searches. Anomalies can be detected by measuring the distance between a data point's vector and its nearest neighbors—if a point is significantly distant, it's likely anomalous. This approach is scalable, adaptable to various data types, and effective for tasks like fraud detection, predictive maintenance, and cybersecurity.
# Anomaly Detection

@ -1,7 +1 @@
# Anthropic's Claude
Claude is a family of large language models developed by Anthropic. Claude 3.5 Sonnet is the latest model (at the time of this writing) in the series, known for its advanced reasoning and multi-modality capabilities.
Visit the following resources to learn more:
- [@official@Clause Website](https://claude.ai/)
# Anthropic's Claude

@ -1,3 +1 @@
# Audio Processing
Using Multimodal AI, audio data can be processed with other types of data, such as text, images, or video, to enhance understanding and analysis. For example, it can synchronize audio with corresponding visual inputs, like lip movements in video, to improve speech recognition or emotion detection. This fusion of modalities enables more accurate transcription, better sentiment analysis, and enriched context understanding in applications such as virtual assistants, multimedia content analysis, and real-time communication systems.
# Audio Processing

@ -1,7 +1 @@
# AWS Sagemaker
AWS Sagemaker is a fully managed platform that provides every developer and data scientist with the ability to build, train, and deploy machine learning (ML) models quickly. Sagemaker takes care of the underlying infrastructure, allowing developers to focus on building and improving their models.
Visit the following resources to learn more:
- [@official@AWS Website](https://aws.amazon.com/sagemaker/)
# AWS Sagemaker

@ -1,7 +1 @@
# Azure AI
Azure AI is a comprehensive set of AI services and tools provided by Microsoft. It includes a range of capabilities such as natural language processing, computer vision, speech recognition, and more. Azure AI is designed to help developers and organizations build, deploy, and scale AI solutions quickly and easily.
Visit the following resources to learn more:
- [@official@Azure Website](https://azure.microsoft.com/en-us/products/ai-services/)
# Azure AI

@ -1,7 +1 @@
# Benefits of Pre-trained Models
LLM models are not only difficult to train, but they are also expensive. Pre-trained models are a cost-effective solution for developers and organizations looking to leverage the power of AI without the need to train models from scratch.
Visit the following resources to learn more:
- [@article@Why you should use Pre-trained Models](https://cohere.com/blog/pre-trained-vs-in-house-nlp-models)
# Benefits of Pre-trained Models

@ -1,3 +1 @@
# Bias and Fareness
Bias and fairness in AI arise when models produce skewed or unequal outcomes for different groups, often reflecting imbalances in the training data. This can lead to discriminatory effects in critical areas like hiring, lending, and law enforcement. Addressing these concerns involves ensuring diverse and representative data, implementing fairness metrics, and ongoing monitoring to prevent biased outcomes. Techniques like debiasing algorithms and transparency in model development help mitigate bias and promote fairness in AI systems.
# Bias and Fareness

@ -1,7 +1 @@
# Capabilities / Context Length
OpenAI's capabilities include processing complex tasks like language understanding, code generation, and problem-solving. However, context length limits how much information the model can retain and reference during a session, affecting long conversations or documents. Advances aim to increase this context window for more coherent and detailed outputs over extended interactions.
Visit the following resources to learn more:
- [@official@OpenAI Website](https://platform.openai.com/docs/guides/fine-tuning/token-limits)
# Capabilities / Context Length

@ -1,7 +1 @@
# Chat Completions API
The Chat Completions API allows developers to create conversational agents by sending user inputs (prompts) and receiving model-generated responses. It supports multiple-turn dialogues, maintaining context across exchanges to deliver relevant responses. This API is often used for chatbots, customer support, and interactive applications where maintaining conversation flow is essential.
Visit the following resources to learn more:
- [@official@OpenAI Website](https://platform.openai.com/docs/api-reference/chat/completions)
# Chat Completions API

@ -1,7 +1 @@
# Chroma
Chroma is a vector database designed to efficiently store, index, and query high-dimensional embeddings. It’s optimized for AI applications like semantic search, recommendation systems, and anomaly detection by allowing fast retrieval of similar vectors based on distance metrics (e.g., cosine similarity). Chroma enables scalable and real-time processing, making it a popular choice for projects involving embeddings from text, images, or other data types.
Visit the following resources to learn more:
- [@official@Chroma Website](https://docs.trychroma.com/)
# Chroma

@ -1,3 +1 @@
# Chunking
In Retrieval-Augmented Generation (RAG), **chunking** refers to breaking large documents or data into smaller, manageable pieces (chunks) to improve retrieval and generation efficiency. This process helps the system retrieve relevant information more accurately by indexing these chunks in a vector database. During a query, the model retrieves relevant chunks instead of entire documents, which enhances the precision of the generated responses and allows better handling of long-form content within the context length limits.
# Chunking

@ -1,10 +1 @@
# Code Completion Tools
AI Code Completion Tools are software tools that use AI models to assist with code generation and editing. These tools help developers write code more quickly and efficiently by providing suggestions, completing code snippets, and suggesting improvements. AI Code Completion Tools can also be used to generate documentation, comments, and other code-related content.
Visit the following resources to learn more:
- [@website@GitHub Copilot](https://copilot.github.com/)
- [@website@Codeium](https://codeium.com/)
- [@website@Supermaven](https://supermaven.com/)
- [@website@TabNine](https://www.tabnine.com/)
# Code Completion Tools

@ -1,7 +1 @@
# Cohere
Cohere is an AI platform that provides natural language processing (NLP) models and tools, enabling developers to integrate powerful language understanding capabilities into their applications. It offers features like text generation, semantic search, classification, and embeddings. Cohere focuses on scalability and ease of use, making it popular for tasks such as content creation, customer support automation, and building search engines with advanced semantic understanding. It also provides a user-friendly API for custom NLP applications.
Visit the following resources to learn more:
- [@website@Cohere](https://cohere.com/)
# Cohere

@ -1,3 +1 @@
# Conducting Adversarial Testing
Adversarial testing involves creating malicious inputs to test the robustness of AI models. This includes testing for prompt injection, evasion, and other adversarial attacks.
# Conducting adversarial testing

@ -1,3 +1 @@
# Constraining Outputs and Inputs
Constraining outputs and inputs is important for controlling the behavior of AI models. This includes techniques like output filtering, input validation, and rate limiting.
# Constraining outputs and inputs

@ -1,3 +1 @@
# Cut-off Dates / Knowledge
OpenAI models have a knowledge cutoff date, meaning they only have access to information available up until a specific time. For example, my knowledge is up to date until September 2023. As a result, I may not be aware of recent developments, events, or newly released technology. Additionally, these models don’t have real-time internet access, so they can't retrieve or update information beyond their training data. This can limit the ability to provide the latest details or react to rapidly changing topics.
# Cut-off Dates / Knowledge

@ -1,3 +1 @@
# DALL-E API
The DALL-E API allows developers to integrate OpenAI's image generation model into their applications. Using text-based prompts, the API generates unique images that match the descriptions provided by users. This makes it useful for tasks like creative design, marketing, product prototyping, and content creation. The API is highly customizable, enabling developers to adjust parameters such as image size and style. DALL-E excels at creating visually rich content from textual descriptions, expanding the possibilities for AI-driven creative workflows.
# DALL-E API

@ -1,3 +1 @@
# Data Classification
Embeddings are used in data classification by converting data (like text or images) into numerical vectors that capture underlying patterns and relationships. These vector representations make it easier for machine learning models to distinguish between different classes based on the similarity or distance between vectors in high-dimensional space. By training a classifier on these embeddings, tasks like sentiment analysis, document categorization, and image classification can be performed more accurately and efficiently. Embeddings simplify complex data and enhance classification by highlighting key features relevant to each class.
# Data Classification

@ -1,3 +1 @@
# Development Tools
A lot of developer related tools have popped up since the AI revolution. It's being used in the coding editors, in the terminal, in the CI/CD pipelines, and more.
# Development Tools

@ -1,3 +1 @@
# Embedding
Embedding refers to the conversion or mapping of discrete objects such as words, phrases, or even entire sentences into vectors of real numbers. It's an essential part of data preprocessing where high-dimensional data is transformed into a lower-dimensional equivalent. This dimensional reduction helps to preserve the semantic relationships between objects. In AI engineering, embedding techniques are often used in language-orientated tasks like sentiment analysis, text classification, and Natural Language Processing (NLP) to provide an understanding of the vast linguistic inputs AI models receive.
# Embedding

@ -1,3 +1 @@
# Embedding
Embedding refers to the conversion or mapping of discrete objects such as words, phrases, or even entire sentences into vectors of real numbers. It's an essential part of data preprocessing where high-dimensional data is transformed into a lower-dimensional equivalent. This dimensional reduction helps to preserve the semantic relationships between objects. In AI engineering, embedding techniques are often used in language-orientated tasks like sentiment analysis, text classification, and Natural Language Processing (NLP) to provide an understanding of the vast linguistic inputs AI models receive.
# Embeddings

@ -1,3 +1 @@
# FAISS
FAISS stands for Facebook AI Similarity Search, it is a database management library developed by Facebook's AI team. Primarily used for efficient similarity search and clustering of dense vectors, it allows users to search through billions of feature vectors swiftly and efficiently. As an AI engineer, learning FAISS is beneficial because these vectors represent objects that are typically used in machine learning or AI applications. For instance, in an image recognition task, a dense vector might be a list of pixels from an image, and FAISS allows a quick search of similar images in a large database.
# FAISS

@ -1,7 +1 @@
# Fine-tuning
OpenAI API allows you to fine-tune and adapt pre-trained models to specific tasks or datasets, improving performance on domain-specific problems. By providing custom training data, the model learns from examples relevant to the intended application, such as specialized customer support, unique content generation, or industry-specific tasks.
Visit the following resources to learn more:
- [@official@OpenAI Docs](https://platform.openai.com/docs/guides/fine-tuning)
# Fine-tuning

@ -1,3 +1 @@
# Generation
In this step of implementing RAG, we use the found chunks to generate a response to the user's query using an LLM.
# Generation

@ -1,3 +1 @@
# Google's Gemini
Gemini, formerly known as Bard, is a generative artificial intelligence chatbot developed by Google. Based on the large language model of the same name, it was launched in 2023 after being developed as a direct response to the rise of OpenAI's ChatGPT
# Google's Gemini

@ -1,7 +1 @@
# Hugging Face Hub
Hugging Face Hub is a platform where you can share, access and collaborate upon a wide array of machine learning models, primarily focused on Natural Language Processing (NLP) tasks. It is a central repository that facilitates storage and sharing of models, reducing the time and overhead usually associated with these tasks. For an AI Engineer, leveraging Hugging Face Hub can accelerate model development and deployment, effectively allowing them to work on structuring efficient AI solutions instead of worrying about model storage and accessibility issues.
Visit the following resources to learn more:
- [@official@Hugging Face](https://huggingface.co/)
# Hugging Face Hub

@ -1,7 +1 @@
# Hugging Face Models
Hugging Face has a wide range of pre-trained models that can be used for a variety of tasks, including language understanding and generation, translation, chatbots, and more. Anyone can create an account and use their models, and the models are organized by task, provider, and other criteria.
Visit the following resources to learn more:
- [@official@Hugging Face](https://huggingface.co/models)
# Hugging Face Models

@ -1,7 +1 @@
# Hugging Face Models
Hugging Face has a wide range of pre-trained models that can be used for a variety of tasks, including language understanding and generation, translation, chatbots, and more. Anyone can create an account and use their models, and the models are organized by task, provider, and other criteria.
Visit the following resources to learn more:
- [@official@Hugging Face](https://huggingface.co/models)
# Hugging Face Models

@ -1,7 +1 @@
# Hugging Face Tasks
Hugging face has a section where they have a list of tasks with the popular models for that task.
Visit the following resources to learn more:
- [@official@Hugging Face](https://huggingface.co/tasks)
# Hugging Face Tasks

@ -1,7 +1 @@
# Hugging Face
Hugging Face is the platform where the machine learning community collaborates on models, datasets, and applications.
Visit the following resources to learn more:
- [@official@Hugging Face](https://huggingface.co/)
# Hugging Face

@ -1,3 +1 @@
# Image Generation
Image Generation often refers to the process of creating new images from an existing dataset or completely from scratch. For an AI Engineer, understanding image generation is crucial as it is one of the key aspects of machine learning and deep learning related to computer vision. It often involves techniques like convolutional neural networks (CNN), generative adversarial networks (GANs), and autoencoders. These technologies are used to generate artificial images that closely resemble original input, and can be applied in various fields such as healthcare, entertainment, security and more.
# Image Generation

@ -1,3 +1 @@
# Image Understanding
Image Understanding involves extracting meaningful information from images, such as photos or videos. This process includes tasks like image recognition, where an AI system is trained to recognize certain objects within an image, and image segmentation, where an image is divided into multiple regions according to some criteria. For an AI engineer, mastering techniques in Image Understanding is crucial because it forms the basis for more complex tasks such as object detection, facial recognition, or even whole scene understanding, all of which play significant roles in various AI applications. As AI technologies continue evolving, the ability to analyze and interpret visual data becomes increasingly important in fields ranging from healthcare to autonomous vehicles.
# Image Understanding

@ -1,3 +1 @@
# Impact on Product Development
Incorporating Artificial Intelligence (AI) can transform the process of creating, testing, and delivering products. This could range from utilizing AI for enhanced data analysis to inform product design, use of AI-powered automation in production processes, or even AI as a core feature of the product itself.
# Impact on Product Development

@ -1,3 +1 @@
# Indexing Embeddings
This step involves converting data (such as text, images, or other content) into numerical vectors (embeddings) using a pre-trained model. These embeddings capture the semantic relationships between data points. Once generated, the embeddings are stored in a vector database, which organizes them in a way that enables efficient retrieval based on similarity. This indexed structure allows fast querying and comparison of vectors, facilitating tasks like semantic search, recommendation systems, and anomaly detection.
# Indexing Embeddings

@ -1,8 +1 @@
# Inference SDK
Inference is the process of using a trained model to make predictions on new data. As this process can be compute-intensive, running on a dedicated server can be an interesting option. The huggingface_hub library provides an easy way to call a service that runs inference for hosted models. There are several services you can connect to:
Visit the following resources to learn more:
- [@official@Hugging Face Inference Client](https://huggingface.co/docs/huggingface_hub/en/package_reference/inference_client)
- [@official@Hugging Face Inference API](https://huggingface.co/docs/api-inference/en/index)
# Inference SDK

@ -1,3 +1 @@
# Inference
Inference involves using models developed through machine learning to make predictions or decisions. As part of the AI Engineer Roadmap, an AI engineer might create an inference engine, which uses rules and logic to infer new information based on existing data. Often used in natural language processing, image recognition, and similar tasks, inference can help AI systems provide useful outputs based on their training. Working with inference involves understanding different models, how they work, and how to apply them to new data to achieve reliable results.
# Inference

@ -1,3 +1 @@
# Introduction
An AI Engineer uses pre-trained models and existing AI tools to improve user experiences. They focus on applying AI in practical ways, without building models from scratch. This is different from AI Researchers and ML Engineers, who focus more on creating new models or developing AI theory.
# Introduction

@ -1,3 +1 @@
# Know your Customers / Usecases
Understanding your target customers and use-cases helps making informed decisions during the development to ensure that the final AI solution appropriately meets the relevant needs of the users. You can use this knowledge to choose the right tools, frameworks, technologies, design the right architecture, and even prevent abuse.
# Know your Customers / Usecases

@ -1,3 +1 @@
# LanceDB
LanceDB is a relatively new, multithreaded, high-speed data warehouse optimized for AI and machine learning data processing. It's designed to handle massive amounts of data, enables quick storage and retrieval, and supports lossless data compression. For an AI engineer, learning LanceDB could be beneficial as it can be integrated with machine learning frameworks for collecting, processing and analyzing large datasets. These functionalities can help to streamline the process for AI model training, which requires extensive data testing and validation.
# LanceDB

@ -1,3 +1 @@
# LangChain for Multimodal Apps
LangChain is a software framework that helps facilitate the integration of large language models into applications. As a language model integration framework, LangChain's use-cases largely overlap with those of language models in general, including document analysis and summarization, chatbots, and code analysis.
# LangChain for Multimodal Apps

@ -1,3 +1 @@
# Langchain
LangChain is a software framework that helps facilitate the integration of large language models into applications. As a language model integration framework, LangChain's use-cases largely overlap with those of language models in general, including document analysis and summarization, chatbots, and code analysis.
# Langchain

@ -1,3 +1 @@
# Limitations and Considerations under Pre-trained Models
Pre-trained Models are AI models that are previously trained on a large benchmark dataset and provide a starting point for AI developers. They help in saving training time and computational resources. However, they also come with certain limitations and considerations. These models can sometimes fail to generalize well to tasks outside of their original context due to issues like dataset bias or overfitting. Furthermore, using them without understanding their internal working can lead to problematic consequences. Finally, transfer learning, which is the mechanism to deploy these pre-trained models, might not always be the optimum solution for every AI project. Thus, an AI Engineer must be aware of these factors while working with pre-trained models.
# Limitations and Considerations

@ -1,7 +1 @@
# Llama Index
LlamaIndex is a simple, flexible data framework for connecting custom data sources to large language models.
Visit the following resources to learn more:
- [@official@LlamaIndex Official Website](https://llamaindex.ai/)
# Llama Index

@ -1,7 +1 @@
# Llama Index
LlamaIndex is a simple, flexible data framework for connecting custom data sources to large language models.
Visit the following resources to learn more:
- [@official@LlamaIndex Official Website](https://llamaindex.ai/)
# LlamaIndex for Multimodal Apps

@ -1,3 +1 @@
# LLMs
LLM or Large Language Models are AI models that are trained on a large amount of text data to understand and generate human language. They are the core of applications like ChatGPT, and are used for a variety of tasks, including language translation, question answering, and more.
# LLMs

@ -1,3 +1 @@
# Manual Implementation
You can build the AI agents manually by coding the logic from scratch without using any frameworks or libraries. For example, you can use the OpenAI API and write the looping logic yourself to keep the agent running until it has the answer.
# Manual Implementation

@ -1,9 +1 @@
# Maximum Tokens
Number of Maximum tokens in OpenAI API depends on the model you are using.
For example, the `gpt-4o` model has a maximum of 128,000 tokens.
Visit the following resources to learn more:
- [@official@OpenAI API Documentation](https://platform.openai.com/docs/api-reference/completions/create)
# Maximum Tokens

@ -1,3 +1 @@
# Mistral AI
Mistral AI is a French startup founded in 2023, specializing in open-source large language models (LLMs). Created by former Meta and Google DeepMind researchers, it focuses on efficient, customizable AI solutions that promote transparency. Its flagship models, Mistral Large and Mixtral, offer state-of-the-art performance with lower resource requirements, gaining significant attention in the AI field.
# Mistral AI

@ -1,7 +1 @@
# Hugging Face Models
Hugging Face has a wide range of pre-trained models that can be used for a variety of tasks, including language understanding and generation, translation, chatbots, and more. Anyone can create an account and use their models, and the models are organized by task, provider, and other criteria.
Visit the following resources to learn more:
- [@official@Hugging Face](https://huggingface.co/models)
# Models on Hugging Face

@ -1,7 +1 @@
# MongoDB Atlas
MongoDB Atlas is a fully managed cloud-based NoSQL database service by MongoDB. It simplifies database deployment and management across platforms like AWS, Azure, and Google Cloud. Using a flexible document model, Atlas automates tasks such as scaling, backups, and security, allowing developers to focus on building applications. With features like real-time analytics and global clusters, it offers a powerful solution for scalable and resilient app development.
Visit the following resources to learn more:
- [@official@MongoDB Atlas Vector Search](https://www.mongodb.com/products/platform/atlas-vector-search)
# MongoDB Atlas

@ -1,3 +1 @@
# Multimodal AI Usecases
Multimodal AI integrates various data types for diverse applications. In human-computer interaction, it enhances interfaces using speech, gestures, and facial expressions. In healthcare, it combines medical scans and records for accurate diagnoses. For autonomous vehicles, it processes data from sensors for real-time navigation. Additionally, it generates images from text and summarizes videos in content creation, while also analyzing satellite and sensor data for climate insights.
# Multimodal AI Usecases

@ -1,3 +1 @@
# Multimodal AI
Multimodal AI refers to artificial intelligence systems capable of processing and integrating multiple types of data inputs simultaneously, such as text, images, audio, and video. Unlike traditional AI models that focus on a single data type, multimodal AI combines various inputs to achieve a more comprehensive understanding and generate more robust outputs. This approach mimics human cognition, which naturally integrates information from multiple senses to form a complete perception of the world. By leveraging diverse data sources, multimodal AI can perform complex tasks like image captioning, visual question answering, and cross-modal content generation.
# Multimodal AI

@ -1,7 +1 @@
# Ollama Models
Ollama supports a wide range of language models, including but not limited to Llama, Phi, Mistral, Gemma and more.
Visit the following resources to learn more:
- [@official@Ollama Models](https://ollama.com/library)
# Ollama Models

@ -1,7 +1 @@
# Ollama SDK
Ollama SDK can be used to develop applications locally.
Visit the following resources to learn more:
- [@official@Ollama SDK](https://ollama.com)
# Ollama SDK

@ -1,7 +1 @@
# Ollama
Ollama is an open-source tool for running large language models (LLMs) locally on personal computers. It supports various models like Llama 2, Mistral, and Code Llama, bundling weights, configurations, and data into a single package. Ollama offers a user-friendly interface, API access, and integration capabilities, allowing users to leverage AI capabilities while maintaining data privacy and control. It's designed for easy installation and use on macOS and Linux, with Windows support in development.
Visit the following resources to learn more:
- [@official@Ollama](https://ollama.com)
# Ollama

@ -1,3 +1 @@
# Open AI Assistant API
OpenAI Assistant API is a tool provided by OpenAI that allows developers to integrate the same AI used in ChatGPT into their own applications, products or services. This AI conducts dynamic, interactive and context-aware conversations useful for building AI assistants in various applications. In the AI Engineer Roadmap, mastering the use of APIs like the Open AI Assistant API is a crucial skill, as it allows engineers to harness the power and versatility of pre-trained algorithms and use them for their desired tasks. AI Engineers can offload the intricacies of model training and maintenance, focusing more on product development and innovation.
# Open AI Assistant API

@ -1,7 +1 @@
# Open AI Embedding Models
Open AI embedding models refer to the artificial intelligence variants designed to reformat or transcribe input data into compact, dense numerical vectors. These models simplify and reduce the input data from its original complex nature, creating a digital representation that is easier to manipulate. This data reduction technique is critical in the AI Engineer Roadmap because it paves the way for natural language processing tasks. It helps in making precise predictions, clustering similar data, and producing accurate search results based on contextual relevance.
Visit the following resources to learn more:
- [@official@Open AI Embedding Models](https://platform.openai.com/docs/guides/embeddings)
# Open AI Embedding Models

@ -1,3 +1 @@
# Open AI Embeddings API
Open AI Embeddings API is a powerful system that is used to generate high-quality word and sentence embeddings. With this API, it becomes a breeze to convert textual data into a numerical format that Machine Learning models can process. This conversion of text into numerical data is crucial for Natural Language Processing (NLP) tasks that an AI Engineer often encounters. Understanding and harnessing the capabilities of the Open AI Embeddings API, therefore, forms an essential part of the AI Engineer's roadmap.
# Open AI Embeddings API

@ -1,3 +1 @@
# Open AI Models
Open AI Models are a set of pre-designed, predefined models provided by OpenAI. These models are trained using Machine Learning algorithms to perform artificial intelligence tasks without any need of explicit programming. OpenAI's models are suited for various applications such as text generation, classification and extraction, allowing AI engineers to leverage them for effective implementations. Therefore, understanding and utilizing these models becomes an essential aspect in the roadmap for an AI engineer to develop AI-powered solutions with more efficiency and quality.
# Open AI Models

@ -1,3 +1 @@
# Open AI Playground
Open AI Playground is an interactive platform, provided by OpenAI, that enables developers to experiment with and understand the capabilities of OpenAI's offerings. Here, you can try out several cutting-edge language models like GPT-3 or Codex. This tool is crucial in the journey of becoming an AI Engineer, because it provides a hands-on experience in implementing and testing language models. Manipulating models directly helps you get a good grasp on how AI models can influence the results based on input parameters. Therefore, Open AI Playground holds significance on the AI Engineer's roadmap not only as a learning tool, but also as a vital platform for rapid prototyping and debugging.
# Open AI Playground

@ -1,3 +1 @@
# Open-Source Embeddings
Open-source embeddings, such as Word2Vec, GloVe, and FastText, are essentially vector representations of words or phrases. These representations capture the semantic relationships between words and their surrounding context in a multi-dimensional space, making it easier for machine learning models to understand and process textual data. In the AI Engineer Roadmap, gaining knowledge of open-source embeddings is critical. These embeddings serve as a foundation for natural language processing tasks, ranging from sentiment analysis to chatbot development, and are widely used in the AI field for their ability to enhance the performance of machine learning models dealing with text data.
# Open-Source Embeddings

@ -1,3 +1 @@
# Open vs Closed Source Models
Open source models are types of software whose source code is available for the public to view, modify, and distribute. They encourage collaboration and transparency, often resulting in rapid improvements and innovations. Closed source models, on the other hand, do not make their source code available and are typically developed and maintained by specific companies or teams. They often provide more stability, support, and consistency. Within the AI Engineer Roadmap, both open and closed source models play a unique role. While open source models allow for customization, experimentation and a broader understanding of underlying algorithms, closed source models might offer proprietary algorithms and structures that could lead to more efficient or unique solutions. Therefore, understanding the differences, advantages, and drawbacks of both models is essential for an aspiring AI engineer.
# Open vs Closed Source Models

@ -1,3 +1 @@
# OpenAI API
OpenAI API is a powerful language model developed by OpenAI, a non-profit artificial intelligence research organization. It uses machine learning to generate text from a given set of keywords or sentences, presenting the capability to learn, understand, and generate human-friendly content. As an AI Engineering aspirant, familiarity with tools like the OpenAI API positions you on the right track. It can help with creating AI applications that can analyze and generate text, which is particularly useful in AI tasks such as data extraction, summarization, translation, and natural language processing.
# OpenAI API

@ -1,3 +1 @@
# OpenAI Assistant API
OpenAI Assistant API is a tool developed by OpenAI which allows developers to establish interaction between their applications, products or services and state-of-the-art AI models. By integrating this API in their software architecture, artificial intelligence engineers can leverage the power of advanced language models developed by the OpenAI community. These integrated models can accomplish a multitude of tasks, like writing emails, generating code, answering questions, tutoring in different subjects and even creating conversational agents. For an AI engineer, mastery over such APIs means they can deploy and control highly sophisticated AI models with just a few lines of code.
# OpenAI Assistant API

@ -1,3 +1 @@
# OpenAI Functions / Tools
OpenAI, a leading organization in the field of artificial intelligence, provides a suite of functions and tools to enable developers and AI engineers to design, test, and deploy AI models. These tools include robust APIs for tasks like natural language processing, vision, and reinforcement learning, and platforms like GPT-3, CLIP, and Codex that provide pre-trained models. Utilization of these OpenAI components allows AI engineers to get a head-start in application development, simplifying the process of integration and reducing the time required for model training and tuning. Understanding and being adept at these tools forms a crucial part of the AI Engineer's roadmap to build impactful AI-driven applications.
# OpenAI Functions / Tools

@ -1,3 +1 @@
# OpenAI Models
OpenAI is an artificial intelligence research lab that is known for its cutting-edge models. These models, like GPT-3, are pre-trained on vast amounts of data and perform remarkably well on tasks like language translation, question-answering, and more without needing any specific task training. Using these pre-trained models can give a massive head-start in building AI applications, as it saves the substantial time and resources that are required for training models from scratch. For an AI Engineer, understanding and leveraging these pre-trained models can greatly accelerate development and lead to superior AI systems.
# OpenAI Models

@ -1,3 +1 @@
# OpenAI Moderation API
OpenAI Moderation API is a feature or service provided by OpenAI that helps in controlling or filtering the output generated by an AI model. It is highly useful in identifying and preventing content that violates OpenAI’s usage policies from being shown. As an AI engineer, learning to work with this API helps implement a layer of security to ensure that the AI models developed are producing content that aligns with the ethical and moral guidelines set in place. Thus, it becomes a fundamental aspect of the AI Engineer Roadmap when dealing with user-generated content or creating AI-based services that interact with people.
# OpenAI Moderation API

@ -1,3 +1 @@
# OpenAI Vision API
OpenAI Vision API is an API provided by OpenAI that is designed to analyze and generate insights from images. By feeding it an image, the Vision API can provide information about the objects and activities present in the image. For AI Engineers, this tool can be particularly useful for conducting Computer Vision tasks effortlessly. Using this API can support in creating applications that need image recognition, object detection and similar functionality, saving AI Engineers from having to create complex image processing algorithms from scratch. Understanding how to work with APIs, especially ones as advanced as the OpenAI Vision API, is an essential skill in the AI Engineer's roadmap.
# OpenAI Vision API

@ -1,3 +1 @@
# OpenSource AI
OpenSource AI refers to artificial intelligence tools, software, libraries and platforms that are freely available to the public, allowing individuals and organizations to use, modify and distribute them as per their requirements. The OpenSource AI initiatives provide an ecosystem for AI developers to innovate, collaborate and mutually learn by sharing their codebase and datasets. Specifically, in the AI engineer's roadmap, OpenSource AI aids in accelerating the AI application development process, provides access to pre-trained models, and promotes the understanding of AI technology through transparency.
# OpenSource AI

@ -1,3 +1 @@
# Performing Similarity Search
This step involves querying the vector database to find the most similar embeddings to a given input vector. When a query is made, the system computes the distance between the input vector and stored embeddings using metrics like cosine similarity or Euclidean distance. The closest matches—those with the smallest distances—are retrieved as results, allowing for accurate semantic search, recommendations, or content retrieval based on similarity in the embedded space. This process enables highly efficient and relevant searches across large datasets.
# Performing Similarity Search

@ -1,3 +1 @@
# Pinecone
Pinecone is a vector database designed specifically for machine learning applications. It facilitates the process of transforming data into a vector and indexing it for quick retrieval. As a cloud-based service, it allows AI Engineers to easily handle high-dimensional data and utilize it for building models. As part of an AI Engineer's Roadmap, understanding and using vector databases like Pinecone can help streamline the development and deployment of AI and ML applications. This is particularly useful in building recommendation systems, personalized search and similarity search which are important components of an AI-based service.
# Pinecone

@ -1,3 +1 @@
# Popular Open Source Models in AI
Open-source models consist of pre-made algorithms and mathematical models that are freely available for anyone to use, modify, and distribute. In the realm of Artificial Intelligence, these models often include frameworks for machine learning, deep learning, natural language processing, and other AI methodologies. Thanks to their openly accessible nature, AI engineers often utilize these open-source models during project execution, fostering increased efficiency by reducing the need to create complex models from scratch. They serve as a valuable resource, often speeding up the development phase and promoting collaboration among the global AI community. Popular examples include TensorFlow, PyTorch, and Keras, each offering unique strengths and capabilities for different areas of AI engineering.
# Popular Open Source Models

@ -1,3 +1 @@
# Pre-trained Models
Pre-trained models are simply models created by some machine learning engineers to solve a problem. Such models are often shared and other machine learning engineers use these models for similar problems. These models are called pre-trained models because they have been previously trained by using large datasets. These pre-trained models can be used as the starting point for a variety of AI tasks, often as part of transfer learning, to save on the resources that would be needed to start a learning process from scratch. This hastens the journey of becoming an AI engineer as one gets to understand how to improve and fine-tune pre-existing models to specific tasks, making them an essential part of an AI engineer's development plan.
# Pre-trained Models

@ -1,3 +1 @@
# Pricing Considerations in OpenAI Embeddings API
OpenAI Embeddings API allows users to compute and extract textual embeddings from large-scale models that OpenAI trains. The pricing for this API can vary based on multiple factors like the number of requests, number of tokens in the text, total computation time, throughput, and others. Understanding the pricing model for the OpenAI Embeddings API is vital for AI Engineers to effectively manage costs while using the API. They should be aware of any limitations or additional costs associated with high volume requests, speed of processing, or special features they plan to use. This knowledge helps the engineers to optimize costs, which is important for the budgeting of AI projects and the overall roadmap of an AI Engineer.
# Pricing Considerations

@ -1,3 +1 @@
# Pricing Considerations
Pricing Considerations refer to the factors and elements that need to be taken into account when setting the price for a product or service. It includes aspects such as cost of production, market demand, competition, and perceived value. In the AI Engineer Roadmap, it can denote the determination of the cost involved in AI model development, implementation, maintenance, and upgrades. Various factors such as the complexity of models, the resources required, timeline, expertise needed, and the value provided to the user play a significant role in pricing considerations.
# Pricing Considerations

@ -1,3 +1 @@
# Prompt Engineering
Prompt Engineering refers to the process of carefully designing and shaping queries or prompts to extract specific responses or behaviors from artificial intelligence models. These prompts are often thought of as the gateway to exploiting these AI models and essential tools for machine testing and performance evaluations. They can affect the model's response, making it invaluable to AI Engineers who are developing AI systems and need to test model's reaction and adaptability with diverse prompts.
# Prompt Engineering

@ -1,3 +1 @@
# Prompt Injection Attacks
Prompt Injection Attacks refer to a cyber threat where nefarious actors manipulate or inject malicious codes into the system using various techniques like SQL injection, Cross-Site Scripting (XSS), or Command Injection. This practice aims to exploit a software system's vulnerabilities, allowing unauthorized access to sensitive information. In the AI Engineer Roadmap, understanding these attack types is essential. Knowledge about such attacks can help developers in AI to build robust and secure AI systems that can resist potential threats and ensure system integrity. Better understanding of threat landscape can guide engineers toward implementing additional security measures during the design and development process of AI applications.
# Prompt Injection Attacks

@ -1,3 +1 @@
# Purpose and Functionality
The Purpose and Functionality are fundamental concepts in the AI Engineer Roadmap. To put simply, 'Purpose' refers to the intended goal or desired result that an AI engineer wants to achieve in an AI project. These goals can range from building neural networks to creating self-driving cars. 'Functionality', on the other hand, pertains to the behaviors and actions that an AI program can perform to fulfill its purpose. This could involve machine learning algorithms, language processing techniques, or data analysis methods among others. Understanding the purpose and functionality of an AI project allows an AI engineer to strategically plan, develop, and manage AI systems effectively.
# Purpose and Functionality

@ -1,3 +1 @@
# Qdrant
Qdrant is a high-performance vector similarity search engine with extensive restful API and distributed support, written in Rust. It allows efficiently storing, handling, and retrieving large amounts of vector data. Integrating Qdrant as a part of the AI Engineer's toolkit can drastically improve functionality and efficiency for AI Engineers, as they often work with vectors during data preprocessing, feature extraction, and modeling. Qdrant's flexibility and control over data indexing and query processing make it particularly handy when dealing with large datasets prevalent in AI projects.
# Qdrant

@ -1,3 +1 @@
# RAG & Implementation
RAG (Relation and Graph) is a mechanism used in artificial intelligence that represents the structured relationships existing between different data entities. Programming languages such as Python provide libraries for RAG implementation, making it simpler for AI engineers. In the AI Engineer roadmap, understanding and implementing RAG models can prove beneficial especially while working with AI algorithms that extensively deal with relational and structured data, such as graph-based Deep Learning algorithms, or while creating knowledge graphs in contexts like Natural Language Processing (NLP). Implementing RAG efficiently can lead to more accurate, efficient, and interpretable AI models.
# RAG & Implementation

@ -1,3 +1 @@
# RAG Usecases
Retrieval-Augmented Generation (RAG) is a type of sequence-to-sequence model with documents retrievers in their architecture. This method integrates the power of pre-trained language models and extractive question answering methods to answer any queries with high precision. In the AI Engineer Roadmap, this tool has practical applications, such as enabling machines to provide detailed responses based on large-scale databases instead of generating responses only from a fixed context. This feature is highly beneficial in developing advanced AI systems with extensive knowledge recall capabilities. RAG's use-cases cover areas like customer service chatbots, automated legal assistance, healthcare advice systems, and other areas where comprehensive information retrieval is crucial.
# RAG Usecases

@ -1,3 +1 @@
# RAG vs Fine-tuning
RAG (Retrieval-Augmented Generation) and Fine-tuning are two distinct techniques utilized in Natural Language Processing (NLP). RAG introduces an approach where the model retrieves documents and faqs from a database to enhance the content generation process. It enables more factual accuracy and relevant context in the outputs. On the other hand, Fine-tuning involves modifying a pre-trained Neural Network model on a new or different task. Adjustments are made to the model's parameters to enhance performance on the new task. Typically, an AI engineer might use RAG for tasks requiring contextual understanding and factual accuracy, while implementing fine-tuning techniques to leverage existing pre-trained models for optimizing new tasks and projects.
# RAG vs Fine-tuning

@ -1,3 +1 @@
# RAG (Retrieval-Augmented Generation)
RAG is a paradigm for applying transformer-based generative models in Natural Language Processing tasks. It leverages a hybrid approach, i.e. it combines the capabilities of pre-trained language models and powerful retrieval methods to generate responses. For an AI Engineer, RAG forms an essential part of the NLP (Natural Language Processing) toolkit. This model operates by first retrieving relevant context from a large corpus, and then utilizing this context to generate detailed and contextually rich responses. Its successful application spans across a multitude of NLP tasks including machine translation, dialogue systems, QnA systems, and more. Therefore, RAG is a significant stop on the route to becoming an accomplished AI engineer as it equips them with skills to deal with complex NLP tasks efficiently.
# RAG

@ -1,3 +1 @@
# ReAct Prompting
ReAct prompting is a tactical approach employed in Conversational AI to generate textual responses. It is essentially utilized in scenarios where a chatbot or an AI-generated persona is required to carry on a conversation. This strategy adds a layer of intelligence to the conversation, maneuvering the AI to generate responses that are contextually sensitive and relevant. In an AI Engineer's Roadmap, an understanding of React Prompting becomes significant during the design of AI interaction models. Using this technique, AI Engineers are capable of creating more intuitive, engaging, and user-friendly conversational AI agents.
# ReAct Prompting

@ -1,3 +1 @@
# Recommendation Systems
Recommendation systems are a subclass of information filtering systems that are meant to predict the preferences or ratings that a user would give to a particular item. Broadly speaking, these systems are primarily used in applications where a user receives suggestions for the products or services they might be interested in, such as Netflix's movie recommendations or Amazon's product suggestions. In terms of an AI Engineer Roadmap, building recommendation systems is a fundamental skill, as these systems typically utilize concepts of machine learning and data mining, and their purpose primarily revolves around making predictions based on large volumes of data. These skills make an integral part of AI-related fields like natural language processing, robotics, and deep learning.
# Recommendation Systems

@ -1,3 +1 @@
# Replicate
Replicate is a version-control tool specifically designed for machine learning. It enables effective tracking of experiments, facilitating the comparison of different models and parameters. As an AI Engineer, knowing how to use Replicate provides you with the ability to save versions of data and model files, thereby preventing loss of work and confusion. It also contributes to smoother teamwork and collaborations by allowing effective sharing and reproduction of experiments, which is crucial in an AI project life cycle.
# Replicate

@ -1,3 +1 @@
# Retrieval Process
In this step of implementing RAG, we clean up the user's query by removing any extra information, we then generate an embedding for the query and look for the most similar embeddings in the vector database.
# Retrieval Process

@ -1,3 +1 @@
# Robust Prompt Engineering
Robust prompt engineering refers to designing, refining, and optimizing the instructions or queries given to an AI model to execute specific tasks. Originally, AI models were trained on a wide range of internet text without any specific commands. However, it can be more effective to provide these models with explicit prompts to guide their responses or actions. Prompt engineering aids in shaping the output of an AI model, significantly improving the accuracy of its responses. This becomes particularly valuable for AI engineers when working on state-of-the-art models like GPT-3, where the output's quality and relevance can be heavily influenced by innovative and well-structured prompts. With robust prompt engineering, AI practitioners can better channel the model's raw capabilities into desired outcomes, marking a crucial skill in an AI Engineer's journey.
# Robust prompt engineering

@ -1,3 +1 @@
# Roles and Responsbilities of an AI Engineer
An AI Engineer is entrusted with the task of designing and implementing AI models. This involves working closely with Data Scientists to transform machine learning models into APIs, ensuring that the models are equipped to interact with software applications. AI Engineers are proficient in a variety of programming languages, work with vast datasets, and utilize AI-related applications. Additionally, they often handle tasks such as data preprocessing, data analysis, and machine learning algorithm deployment. They also troubleshoot any issues that might emerge during the AI lifecycle, while maintaining a high level of knowledge about the latest industry trends and technological advancements.
# Roles and Responsiblities

@ -1,3 +1 @@
# Security and Privacy Concerns
Security and Privacy Concerns encapsulates the understanding and addressing of potential risks associated with AI systems. These include, but are not limited to, data protection, access control, regulatory compliance, and the ethically complex area of how AI impacts individual privacy. As an aspiring AI Engineer, it is essential to acknowledge these concerns alongside technical skills as they influence the design, implementation, and application of AI technologies. Familiarity with this area helps in designing AI solutions that align with standards of security and privacy while effectively addressing the needs of the user.
# Security and Privacy Concerns

@ -1,3 +1 @@
# Semantic Search
Semantic Search is an information retrieval approach which leverages not just the keywords in a search query, but also the intent and contextual meaning behind them to produce highly relevant results. In other words, it makes search engines more intelligent and precise, understanding user intent and making connections like a human brain would. It's an important technique that an AI Engineer might utilize, especially when dealing with large amounts of data or if they're involved in creating intelligent search systems. From natural language processing to relationship mapping, semantic search plays a key role in advancing artificial intelligence search capabilities.
# Semantic Search

@ -1,3 +1 @@
# Sentence Transformers
Sentence Transformers refer to a variant of the popular Transformers model that is specifically designed and optimized for creating meaningful and effective sentence embeddings. It enables developers to easily convert paragraphs and sentences into dense vector representations that can be compared for semantic similarity. In the AI engineer's journey, getting familiar with Sentence Transformers is important because it allows the modelling of natural language in AI systems to provide richer, more nuanced interactions. This can be especially valuable in designing and implementing AI applications such as chatbots, sentiment analysis tools, and more.
# Sentence Transformers

Some files were not shown because too many files have changed in this diff Show More

Loading…
Cancel
Save