<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Tobias Macey</title>
    <description>The latest articles on DEV Community by Tobias Macey (@blarghmatey).</description>
    <link>https://dev.to/blarghmatey</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F115291%2F28dc23f8-4e70-4f75-b666-8b1d5e399484.png</url>
      <title>DEV Community: Tobias Macey</title>
      <link>https://dev.to/blarghmatey</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/blarghmatey"/>
    <language>en</language>
    <item>
      <title>Streaming Data Integration Without The Code at Equalum - Episode 161</title>
      <dc:creator>Tobias Macey</dc:creator>
      <pubDate>Tue, 01 Dec 2020 13:13:20 +0000</pubDate>
      <link>https://dev.to/blarghmatey/streaming-data-integration-without-the-code-at-equalum-episode-161-2cja</link>
      <guid>https://dev.to/blarghmatey/streaming-data-integration-without-the-code-at-equalum-episode-161-2cja</guid>
      <description>&lt;p&gt;The first stage of every good pipeline is to perform data integration. With the increasing pace of change and the need for up to date analytics the need to integrate that data in near real time is growing. With the improvements and increased variety of options for streaming data engines and improved tools for change data capture it is possible for data teams to make that goal a reality. However, despite all of the tools and managed distributions of those streaming engines it is still a challenge to build a robust and reliable pipeline for streaming data integration, especially if you need to expose those capabilities to non-engineers. In this episode Ido Friedman, CTO of Equalum, explains how they have built a no-code platform to make integration of streaming data and change data capture feeds easier to manage. He discusses the challenges that are inherent in the current state of CDC technologies, how they have architected their system to integrate well with existing data platforms, and how to build an appropriate level of abstraction for such a complex problem domain. If you are struggling with streaming data integration and change data capture then this interview is definitely worth a listen.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.dataengineeringpodcast.com/equalum-streaming-data-integration-episode-161/"&gt;Listen Now!&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Open Sourcing The Anvil Full Stack Python Web App Platform - Episode 291</title>
      <dc:creator>Tobias Macey</dc:creator>
      <pubDate>Tue, 01 Dec 2020 13:12:24 +0000</pubDate>
      <link>https://dev.to/blarghmatey/open-sourcing-the-anvil-full-stack-python-web-app-platform-episode-291-34j1</link>
      <guid>https://dev.to/blarghmatey/open-sourcing-the-anvil-full-stack-python-web-app-platform-episode-291-34j1</guid>
      <description>&lt;p&gt;Building a complete web application requires expertise in a wide range of disciplines. As a result it is often the work of a whole team of engineers to get a new project from idea to production. Meredydd Luff and his co-founder built the Anvil platform to make it possible to build full stack applications entirely in Python. In this episode he explains why they released the application server as open source, how you can use it to run your own projects for free, and why developer tooling is the sweet spot for an open source business model. He also shares his vision for how the end-to-end experience of building for the web should look, and some of the innovative projects and companies that were made possible by the reduced friction that the Anvil platform provides. Give it a listen today to gain some perspective on what it could be like to build a web app.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.pythonpodcast.com/anvil-open-source-web-app-server-episode-291/"&gt;Listen Now!&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Self Service Data Management From Ingest To Insights With Isima - Episode 159</title>
      <dc:creator>Tobias Macey</dc:creator>
      <pubDate>Wed, 18 Nov 2020 16:36:57 +0000</pubDate>
      <link>https://dev.to/blarghmatey/self-service-data-management-from-ingest-to-insights-with-isima-episode-159-g7l</link>
      <guid>https://dev.to/blarghmatey/self-service-data-management-from-ingest-to-insights-with-isima-episode-159-g7l</guid>
      <description>&lt;p&gt;The core mission of data engineers is to provide the business with a way to ask and answer questions of their data. This often takes the form of business intelligence dashboards, machine learning models, or APIs on top of a cleaned and curated data set. Despite the rapid progression of impressive tools and products built to fulfill this mission, it is still an uphill battle to tie everything together into a cohesive and reliable platform. At Isima they decided to reimagine the entire ecosystem from the ground up and built a single unified platform to allow end-to-end self service workflows from data ingestion through to analysis. In this episode CEO and co-founder of Isima Darshan Rawal explains how the biOS platform is architected to enable ease of use, the challenges that were involved in building an entirely new system from scratch, and how it can integrate with the rest of your data platform to allow for incremental adoption. This was an interesting and contrarian take on the current state of the data management industry and is worth a listen to gain some additional perspective.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.dataengineeringpodcast.com/isima-data-management-platform-episode-159/"&gt;Listen Now!&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Scale Your Data Science Teams With Machine Learning Operations Principles - Episode 289</title>
      <dc:creator>Tobias Macey</dc:creator>
      <pubDate>Wed, 18 Nov 2020 15:41:47 +0000</pubDate>
      <link>https://dev.to/blarghmatey/scale-your-data-science-teams-with-machine-learning-operations-principles-episode-289-30m2</link>
      <guid>https://dev.to/blarghmatey/scale-your-data-science-teams-with-machine-learning-operations-principles-episode-289-30m2</guid>
      <description>&lt;p&gt;Building a machine learning model is a process that requires well curated and cleaned data and a lot of experimentation. Doing it repeatably and at scale with a team requires a way to share your discoveries with your teammates. This has led to a new set of operational ML platforms. In this episode Michael Del Balso shares the lessons that he learned from building the platform at Uber for putting machine learning into production. He also explains how the feature store is becoming the core abstraction for data teams to collaborate on building machine learning models. If you are struggling to get your models into production, or scale your data science throughput, then this interview is worth a listen.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.pythonpodcast.com/machine-learning-operations-episode-289/"&gt;Listen Now!&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Building A Cost Effective Data Catalog With Tree Schema - Episode 158</title>
      <dc:creator>Tobias Macey</dc:creator>
      <pubDate>Tue, 10 Nov 2020 12:18:08 +0000</pubDate>
      <link>https://dev.to/blarghmatey/building-a-cost-effective-data-catalog-with-tree-schema-episode-158-2i8i</link>
      <guid>https://dev.to/blarghmatey/building-a-cost-effective-data-catalog-with-tree-schema-episode-158-2i8i</guid>
      <description>&lt;p&gt;A data catalog is a critical piece of infrastructure for any organization who wants to build analytics products, whether internal or external. While there are a number of platforms available for building that catalog, many of them are either difficult to deploy and integrate, or expensive to use at scale. In this episode Grant Seward explains how he built Tree Schema to be an easy to use and cost effective option for organizations to build their data catalogs. He also shares the internal architecture, how he approached the design to make it accessible and easy to use, and how it autodiscovers the schemas and metadata for your source systems.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.dataengineeringpodcast.com/tree-schema-data-catalog-episode-158/"&gt;Listen Now!&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Making The Case For A (Semi) Formal Specification Of CPython - Episode 288</title>
      <dc:creator>Tobias Macey</dc:creator>
      <pubDate>Tue, 10 Nov 2020 12:17:33 +0000</pubDate>
      <link>https://dev.to/blarghmatey/making-the-case-for-a-semi-formal-specification-of-cpython-episode-288-1bkm</link>
      <guid>https://dev.to/blarghmatey/making-the-case-for-a-semi-formal-specification-of-cpython-episode-288-1bkm</guid>
      <description>&lt;p&gt;The CPython implementation has grown and evolved significantly over the past ~25 years. In that time there have been many other projects to create compatible runtimes for your Python code. One of the challenges for these other projects is the lack of a fully documented specification of how and why everything works the way that it does. In the most recent Python language summit Mark Shannon proposed implementing a formal specification for CPython, and in this episode he shares his reasoning for why that would be helpful and what is involved in making it a reality.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.pythonpodcast.com/cpython-formal-specification-episode-288/"&gt;Listen Now!&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Cloud Native Data Security As Code With Cyral - Episode 156</title>
      <dc:creator>Tobias Macey</dc:creator>
      <pubDate>Thu, 29 Oct 2020 14:18:48 +0000</pubDate>
      <link>https://dev.to/blarghmatey/cloud-native-data-security-as-code-with-cyral-episode-156-2la2</link>
      <guid>https://dev.to/blarghmatey/cloud-native-data-security-as-code-with-cyral-episode-156-2la2</guid>
      <description>&lt;p&gt;One of the most challenging aspects of building a data platform has nothing to do with pipelines and transformations. If you are putting your workflows into production, then you need to consider how you are going to implement data security, including access controls and auditing. Different databases and storage systems all have their own method of restricting access, and they are not all compatible with each other. In order to simplify the process of securing your data in the Cloud Manav Mital created Cyral to provide a way of enforcing security as code. In this episode he explains how the system is architected, how it can help you enforce compliance, and what is involved in getting it integrated with your existing systems. This was a good conversation about an aspect of data management that is too often left as an afterthought.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.dataengineeringpodcast.com/cyral-data-security-episode-156/"&gt;Listen Now!&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Power Up Your Java Using Python With JPype - Episode 286</title>
      <dc:creator>Tobias Macey</dc:creator>
      <pubDate>Thu, 29 Oct 2020 13:31:51 +0000</pubDate>
      <link>https://dev.to/blarghmatey/power-up-your-java-using-python-with-jpype-episode-286-7ai</link>
      <guid>https://dev.to/blarghmatey/power-up-your-java-using-python-with-jpype-episode-286-7ai</guid>
      <description>&lt;p&gt;Python and Java are two of the most popular programming languages in the world, and have both been around for over 20 years. In that time there have been numerous attempts to provide interoperability between them, with varying methods and levels of success. One such project is JPype, which allows you to use Java classes in your Python code. In this episode the current maintainer, Karl Nelson, explains why he chose it as his preferred tool for combining these ecosystems, how he and his team are using it, and when and how you might want to use it for your own projects. He also discusses the work he has done to enable use of JPype on Android, and what is in store for the future of the project. If you have ever wanted to use a library or module from Java, but the rest of your project is already in Python, then this episode is definitely worth a listen.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.pythonpodcast.com/jpype-java-python-bridge-episode-286/"&gt;Listen Now!&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Better Data Quality Through Observability With Monte Carlo - Episode 155</title>
      <dc:creator>Tobias Macey</dc:creator>
      <pubDate>Tue, 20 Oct 2020 10:36:13 +0000</pubDate>
      <link>https://dev.to/blarghmatey/better-data-quality-through-observability-with-monte-carlo-episode-155-bd</link>
      <guid>https://dev.to/blarghmatey/better-data-quality-through-observability-with-monte-carlo-episode-155-bd</guid>
      <description>&lt;p&gt;In order for analytics and machine learning projects to be useful, they require a high degree of data quality. To ensure that your pipelines are healthy you need a way to make them observable. In this episode Barr Moses and Lior Gavish, co-founders of Monte Carlo, share the leading causes of what they refer to as data downtime and how it manifests. They also discuss methods for gaining visibility into the flow of data through your infrastructure, how to diagnose and prevent potential problems, and what they are building at Monte Carlo to help you maintain your data's uptime.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.dataengineeringpodcast.com/monte-carlo-observability-data-quality-episode-155/"&gt;Listen Now!&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>The Journey To Replace Python's Parser And What It Means For The Future - Episode 285</title>
      <dc:creator>Tobias Macey</dc:creator>
      <pubDate>Tue, 20 Oct 2020 10:35:22 +0000</pubDate>
      <link>https://dev.to/blarghmatey/the-journey-to-replace-python-s-parser-and-what-it-means-for-the-future-episode-285-1lik</link>
      <guid>https://dev.to/blarghmatey/the-journey-to-replace-python-s-parser-and-what-it-means-for-the-future-episode-285-1lik</guid>
      <description>&lt;p&gt;The release of Python 3.9 introduced a new parser that paves the way for brand new features. Every programming language has its own specific syntax for representing the logic that you are trying to express. The way that the rules of the language are defined and validated is with a grammar definition, which in turn is processed by a parser. The parser that the Python language has relied on for the past 25 years has begun to show its age through mounting technical debt and a lack of flexibility in defining new syntax. In this episode Pablo Galindo and Lysandros Nikolaou explain how, together with Python's creator Guido van Rossum, they replaced the original parser implementation with one that is more flexible and maintainable, why now was the time to make the change, and how it will influence the future evolution of the language.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.pythonpodcast.com/cpython-parser-replacement-episode-285/"&gt;Listen Now!&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Rapid Delivery Of Business Intelligence Using Power BI - Episode 154</title>
      <dc:creator>Tobias Macey</dc:creator>
      <pubDate>Mon, 19 Oct 2020 23:10:41 +0000</pubDate>
      <link>https://dev.to/blarghmatey/rapid-delivery-of-business-intelligence-using-power-bi-episode-154-304g</link>
      <guid>https://dev.to/blarghmatey/rapid-delivery-of-business-intelligence-using-power-bi-episode-154-304g</guid>
      <description>&lt;p&gt;Business intelligence efforts are only as useful as the outcomes that they inform. Power BI aims to reduce the time and effort required to go from information to action by providing an interface that encourages rapid iteration. In this episode Rob Collie shares his enthusiasm for the Power BI platform and how it stands out from other options. He explains how he helped to build the platform during his time at Microsoft, and how he continues to support users through his work at Power Pivot Pro. Rob shares some useful insights gained through his consulting work, and why he considers Power BI to be the best option on the market today for business analytics.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.dataengineeringpodcast.com/power-bi-business-intelligence-episode-154/"&gt;Listen Now!&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Cloud Native Application Delivery Using GitOps - Episode 284</title>
      <dc:creator>Tobias Macey</dc:creator>
      <pubDate>Mon, 19 Oct 2020 23:09:52 +0000</pubDate>
      <link>https://dev.to/blarghmatey/cloud-native-application-delivery-using-gitops-episode-284-1knh</link>
      <guid>https://dev.to/blarghmatey/cloud-native-application-delivery-using-gitops-episode-284-1knh</guid>
      <description>&lt;p&gt;The way that applications are being built and delivered has changed dramatically in recent years with the growing trend toward cloud native software. As part of this movement toward the infrastructure and orchestration that powers your project being defined in software, a new approach to operations is gaining prominence. Commonly called GitOps, the main principle is that all of your automation code lives in version control and is executed automatically as changes are merged. In this episode Victor Farcic shares details on how that workflow brings together developers and operations engineers, the challenges that it poses, and how it influences the architecture of your software. This was an interesting look at an emerging pattern in the development and release cycle of modern applications.&lt;/p&gt;
&lt;p&gt;&lt;a href="https://www.pythonpodcast.com/gitops-cloud-native-operations-episode-284/"&gt;Listen Now!&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
