Introduction In today’s rapidly evolving technological landscape, cloud data engineering projects are at the forefront of innovation, driving businesses towards unprecedented levels of efficiency, scalability, and data-driven decision-making. Central to the success of these projects is the concept of distributed teams – groups of individuals who work together from various geographical locations, leveraging the power […]
The final installment of our blog series on AWS testing methodologies focuses on integration testing. This crucial phase ensures that all components of your application work together seamlessly in a live environment, simulating real-world usage with production code and test data. Below is an outline designed to guide the creation of a comprehensive and informative […]
A deep dive into functional testing for AWS development Introduction In our exploration of advanced testing techniques for AWS development, we’ve delved into powerful tools like moto for unit testing and pytest.mark.parametrize for enhancing test coverage and efficiency. Building on this foundation, we turn our focus to a pivotal tool that bridges the gap between […]
Leveraging Moto and Pytest Introduction In the world of AWS development, ensuring the reliability, efficiency, and correctness of your cloud-based applications is paramount. As cloud solutions grow increasingly complex, so too does the challenge of effectively testing these systems. Traditional testing methods often fall short in the face of AWS’s vast and intricately interconnected services. […]
Introduction In the rapidly evolving field of data engineering, maintaining high-quality, reliable, and efficient data pipelines is crucial for businesses to make informed decisions and stay competitive. One methodology that has been instrumental in achieving these objectives is Test-Driven Development (TDD). At its core, TDD involves a simple, yet powerful cycle: write a failing test […]
Generally and historically, data engineering, analytics, and science efforts focused on progressing from data to knowledge/wisdom. The emergence of LLMs allows for the decomposition of wisdom/knowledge back down to data. This can enable novel discovery, integrate with information systems, and drive automated processes. GenAI Categories Generation: Use bedrock models to create code, text, or images […]
Introduction In today’s fast-paced development environment, the Agile methodology stands out for its emphasis on delivering functional features to users as early as possible. This approach challenges traditional, lengthy development cycles by advocating for the incremental release of a product’s most essential functionalities. By prioritizing early delivery, Agile aims to provide immediate value to users, […]
Introduction Joining or starting data projects in large enterprise environments with many stakeholders can be stressful, not to mention a technical implementation nightmare. When the primary stakeholders can’t (or won’t) give the project team clear requirements, the onus falls to the technical implementation team to create order from the chaos and organize the delivery team […]
co-authors: Meghana Venkataswamy, Sean Cahill, Salman Ahmed Mian Architecture What is a Foundational Model? How Do we customize the FM to our Data and needs? Our /hr-bot uses RAG Technique FAISS (Facebook AI Similarity Search) What is AWS Kendra LangChain AWS Bedrock Slack integration Terraform Challenges and opportunities Navigating policy and rules in a large […]
· The role of large language models in creating knowledge graphs from unstructured data. · Comparison of Top Models · Storage Platforms for Knowledge Graphs · Practical Example: Knowledge Graph from Wikipedia Text · Best Practices and Tips for Using Large Language Models in Knowledge Graph Creation · Conclusion In the vast realm of data, […]