<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:media="http://search.yahoo.com/mrss/"><channel><title><![CDATA[Emblica's Blog]]></title><description><![CDATA[Biased samples from our Scientists and Engineers ]]></description><link>https://blog.emblica.fi/</link><image><url>https://blog.emblica.fi/favicon.png</url><title>Emblica&apos;s Blog</title><link>https://blog.emblica.fi/</link></image><generator>Ghost 5.75</generator><lastBuildDate>Mon, 13 Apr 2026 19:13:23 GMT</lastBuildDate><atom:link href="https://blog.emblica.fi/rss/" rel="self" type="application/rss+xml"/><ttl>60</ttl><item><title><![CDATA[Looking Back at 2025: A Year of Meaningful Partnerships in Physical AI]]></title><description><![CDATA[It’s been a year of steady work, clear progress, and a lot of good conversations about where physical AI is headed.]]></description><link>https://blog.emblica.fi/looking-back-at-2025-a-year-of-meaningful-partnerships-in-physical-ai/</link><guid isPermaLink="false">6943c134935c3500010d880c</guid><dc:creator><![CDATA[Teemu Heikkilä]]></dc:creator><pubDate>Thu, 18 Dec 2025 09:15:40 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2025/12/laita-ta-a-.jpg" medium="image"/><content:encoded><![CDATA[<img src="https://blog.emblica.fi/content/images/2025/12/laita-ta-a-.jpg" alt="Looking Back at 2025: A Year of Meaningful Partnerships in Physical AI"><p>As we head into the end of 2025, I&#x2019;ve been reflecting on everything we&#x2019;ve done over the past twelve months. It&#x2019;s been a year of steady work, clear progress, and a lot of good conversations about where physical AI is headed.</p><p>For a long time, we&#x2019;ve focused on making AI work in the real world, not just on a screen, but in heavy machinery and industrial environments. This year, that focus really paid off. We&#x2019;ve moved from just &quot;doing the work&quot; to being a core part of the conversation on how physical AI should be built and used.</p><p><strong>Growing our community</strong></p><p>We&#x2019;ve always believed that technology works better when people collaborate. This year, we got closer to the organizations shaping our industry: We joined <strong>FIMA</strong> (Forum for Intelligent Machines) and were invited to become an associate partner in the <strong>SIX Mobile Work Machines</strong> network. We also continued our work with <strong>AI Finland</strong> and <strong>AHK Finnland</strong>, helping to build a stronger AI ecosystem here and across Europe.</p><p>One thing the team is particularly proud of is our <strong>EmbliCats</strong> initiative. We want the tech world to be more open, so we hosted events like a LiDARLab workshop where we invited women and underrepresented groups in tech to learn about physical AI and LiDARs alongside our experts.</p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2025/12/Teemu_wartsila.jpeg" class="kg-image" alt="Looking Back at 2025: A Year of Meaningful Partnerships in Physical AI" loading="lazy" width="1280" height="854" srcset="https://blog.emblica.fi/content/images/size/w600/2025/12/Teemu_wartsila.jpeg 600w, https://blog.emblica.fi/content/images/size/w1000/2025/12/Teemu_wartsila.jpeg 1000w, https://blog.emblica.fi/content/images/2025/12/Teemu_wartsila.jpeg 1280w" sizes="(min-width: 720px) 720px"></figure><p><strong>Getting out there</strong></p><p>It was great to see so many of you in person this year. Whether we were talking about synthetic data or sharing experiences how does physical AI systems perform in production, these events were the highlight of 2025: </p><p><strong>Bauma 2025:</strong> The world&#x2019;s leading trade fair for construction machinery and equipment.&#xA0;<br><strong>FIMA AI Day: </strong>We co-hosted this event focused on leveraging AI for the R&amp;D of intelligent mobile work machines. We welcomed speakers from Konecranes, Aalto University House of AI, FCAI, and Finnish AI Region FAIR, alongside our own experts.<br><strong>Future Mobile Work Machines 2025: </strong>The flagship event for the mobile machinery sector, where we demonstrated how physical AI systems are built in practice and contributed a keynote.<br><strong>Teknologia 25:</strong> Finland&apos;s leading event for industry and technology professionals. At SICK&#x2019;s booth, we demonstrated real-time human detection and tracking using LiDAR data and a model trained with synthetic data.<br><strong>AI Finland&#x2019;s AI Gala:</strong> Great opportunity to celebrate what the Finnish tech community is achieving together.</p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2025/12/teknologiamessut_emblica--1-.jpg" class="kg-image" alt="Looking Back at 2025: A Year of Meaningful Partnerships in Physical AI" loading="lazy" width="2000" height="1500" srcset="https://blog.emblica.fi/content/images/size/w600/2025/12/teknologiamessut_emblica--1-.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2025/12/teknologiamessut_emblica--1-.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2025/12/teknologiamessut_emblica--1-.jpg 1600w, https://blog.emblica.fi/content/images/size/w2400/2025/12/teknologiamessut_emblica--1-.jpg 2400w" sizes="(min-width: 720px) 720px"></figure><p><strong>Looking toward 2026</strong></p><p>The coming year is looking even busier. We are preparing to grow the team and take our work on more international projects.</p><p>I want to say a sincere thank you to our customers and partners. We don&#x2019;t take your trust for granted, and we&#x2019;re excited to keep building things with you in 2026. If you see us at an event next year, please come over and say hi.</p><p>Wishing you all a great year ahead,</p><p><strong>Teemu Heikkil&#xE4;</strong></p><p>&#xA0;CEO, Emblica</p><hr><p><a href="emblica.com"><em>Emblica</em></a><em> is not your average data science team. We are delivering tailored AI systems for our partners working in unique operation environments and challenges. We design and develop tailored solutions for land, sea, air, and underground applications. By building strong partnerships, we focus on deeply understanding our customers goals, ensuring our development delivers on your vision.</em><br></p>]]></content:encoded></item><item><title><![CDATA[Embeddings and Mushrooms: Another Way of Finding Fungi]]></title><description><![CDATA[Getting your boots dirty is often required when test-driving a new AI concept. This time, I managed to mix that with one of my personal hobbies — mushroom foraging.]]></description><link>https://blog.emblica.fi/embeddings-and-mushrooms/</link><guid isPermaLink="false">68f8c73fe37d3b000140f11c</guid><category><![CDATA[AI]]></category><category><![CDATA[embeddings]]></category><dc:creator><![CDATA[Lotta Koponen]]></dc:creator><pubDate>Tue, 28 Oct 2025 12:29:25 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2025/10/Forest-1.jpg" medium="image"/><content:encoded><![CDATA[<img src="https://blog.emblica.fi/content/images/2025/10/Forest-1.jpg" alt="Embeddings and Mushrooms: Another Way of Finding Fungi"><p>This mini-project turned out to be a great example of how tailored AI systems come to life. While the first goal was simply a tasty dinner, the real takeaway was the power and flexibility of the tech behind it. Similar perceptual systems could easily <strong>scale to autonomous machines</strong> in agriculture, construction, or defense. These systems could enable real-time ground analysis and assess stability or risk in milliseconds to ensure operational safety and efficiency.</p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2025/10/Screenshot-2025-10-28-at-10.41.24.png" class="kg-image" alt="Embeddings and Mushrooms: Another Way of Finding Fungi" loading="lazy" width="1038" height="830" srcset="https://blog.emblica.fi/content/images/size/w600/2025/10/Screenshot-2025-10-28-at-10.41.24.png 600w, https://blog.emblica.fi/content/images/size/w1000/2025/10/Screenshot-2025-10-28-at-10.41.24.png 1000w, https://blog.emblica.fi/content/images/2025/10/Screenshot-2025-10-28-at-10.41.24.png 1038w" sizes="(min-width: 720px) 720px"></figure><hr><p>The idea was simple: build a tool that could learn where the good spots are while we&#x2019;re walking.&#xA0;</p><p>At the core were <strong>embeddings</strong>. They are feature representations created from publicly available Earth Observation data, mainly Sentinel satellite imagery. These embeddings capture a lot of detail about the forest environment (things like soil moisture, canopy density, and terrain structure) at a resolution of about 10&#xD7;10 meters per grid cell. In short, they compress years of observations into a compact description of each location.</p><p>On top of the embeddings we trained a small neural network to predict the probability of finding mushrooms in any given 10&#xD7;10 meter square.</p><p>Then came the fun part &#x2014; we added an <strong>active feedback loop.</strong> While hiking, every time we found a mushroom patch, we marked that grid cell as &#x201C;positive.&#x201D; If we explored an area that looked promising but turned out empty, we marked it as &#x201C;negative.&#x201D; We even logged a few different mushroom species. The model updated itself on the spot, constantly learning from new data. To make that possible, we built a simple web app we could use right from our phones.</p><p>There are already apps out there for mushroom hunting, but most rely on static rules, like biological data and historical sightings. Ours, on the other hand, learns dynamically. It figures out the conditions for growth as they change. Once you find your first patch, it helps guide you to the next one.</p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2025/10/Screenshot-2025-10-28-at-10.19.55.png" class="kg-image" alt="Embeddings and Mushrooms: Another Way of Finding Fungi" loading="lazy" width="1882" height="1002" srcset="https://blog.emblica.fi/content/images/size/w600/2025/10/Screenshot-2025-10-28-at-10.19.55.png 600w, https://blog.emblica.fi/content/images/size/w1000/2025/10/Screenshot-2025-10-28-at-10.19.55.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2025/10/Screenshot-2025-10-28-at-10.19.55.png 1600w, https://blog.emblica.fi/content/images/2025/10/Screenshot-2025-10-28-at-10.19.55.png 1882w" sizes="(min-width: 720px) 720px"></figure><p>After completing these steps, we&#x2019;d trained a model to classify terrain from satellite data just to find mushrooms, but the same input data and the same core capability can be repurposed for much more critical or industrial applications:</p><ul><li><strong>Real-time traversability estimation</strong>: The same features that help locate mushrooms can tell if a forest area is swampy, dense, or uneven. That&#x2019;s invaluable for machines in agriculture, forestry, and construction, helping them plan safe and efficient routes in real time.</li><li><strong>Defense and security</strong>: Terrain &#x201C;walkability&#x201D; is a key factor for operational planning and off-road movement.</li><li><strong>Environmental monitoring</strong>: These classifications can support forestry, conservation, and land management with high precision.</li></ul><hr><p>Our little mushroom seeker project proved that a high-resolution classification layer, trained quickly through active learning, can be incredibly effective. It&#x2019;s a great example of how to build tailored AI &#x2013; we take general models and technologies, combine them with high-quality, context-specific data, and produce a system that gives our clients a unique, practical advantage.</p><p>My biggest takeaway from that weekend? Sometimes, the smartest AI systems start with a simple human touch. A quick loop of feedback and labeling in the field turned a general satellite dataset into a focused, high-resolution solution.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2025/10/Screenshot-2025-10-28-at-10.42.20.png" class="kg-image" alt="Embeddings and Mushrooms: Another Way of Finding Fungi" loading="lazy" width="1074" height="1094" srcset="https://blog.emblica.fi/content/images/size/w600/2025/10/Screenshot-2025-10-28-at-10.42.20.png 600w, https://blog.emblica.fi/content/images/size/w1000/2025/10/Screenshot-2025-10-28-at-10.42.20.png 1000w, https://blog.emblica.fi/content/images/2025/10/Screenshot-2025-10-28-at-10.42.20.png 1074w" sizes="(min-width: 720px) 720px"><figcaption><span style="white-space: pre-wrap;">The model proved to be an excellent mushroom guide. </span></figcaption></figure><p>Want to learn how Emblica can build a cutting-edge AI system for your needs, even if they&#x2019;re as niche as mushroom foraging? <a href="https://www.emblica.com/contact" rel="noreferrer">Let&#x2019;s talk.&#xA0;</a></p><hr><p><a href="https://www.emblica.com/" rel="noreferrer"><em>Emblica</em></a><em>&#xA0;is not your average data team. We build customized solutions for collecting, processing, and utilizing data for all sectors, especially at the R&amp;D interface. Whether our target is a factory line, an online store or a field, you can find us busy at work, hands in the clay - at least at our office in Helsinki.</em></p><hr>]]></content:encoded></item><item><title><![CDATA[Introducing EmbliCats: Empowering Women in Tech at Emblica]]></title><description><![CDATA[On this International Women’s Day, we celebrate not just the achievements of women in technology but also the work being done to create a more supportive, and equal industry. In this blog post, we introduce EmbliCats—our initiative to drive impact on equality and inclusivity in the deep tech sector.]]></description><link>https://blog.emblica.fi/introducing-emblicats-empowering-women-in-tech-at-emblica/</link><guid isPermaLink="false">67caed862603910001f01430</guid><category><![CDATA[Women in tech]]></category><category><![CDATA[EmbliCats]]></category><dc:creator><![CDATA[Sanna Niemelä]]></dc:creator><pubDate>Sat, 08 Mar 2025 07:30:31 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2025/03/kone-2.jpg" medium="image"/><content:encoded><![CDATA[<img src="https://blog.emblica.fi/content/images/2025/03/kone-2.jpg" alt="Introducing EmbliCats: Empowering Women in Tech at Emblica"><p>Women remain underrepresented in the tech industry despite strong evidence that more diverse teams foster innovation, improve decision-making, and boost business performance. In the deep tech environment, where we work with machines, sensors, and advanced data science, the gender gap seems to be even more pronounced compared to the broader IT industry. This is why inclusivity in deep technology is particularly important. At Emblica, we believe that deep tech is for everyone&#x2014;and that&#x2019;s why we&#x2019;re proud to introduce EmbliCats, our internal network dedicated to supporting and uplifting women in tech.</p><p><strong>Who we are</strong></p><p>EmbliCats is a community of women at Emblica, who share a passion for technology, equality, and professional growth. We are proud of the work we do at Emblica, tackling business challenges, already across 13 industries on land, sea, air and underground with cutting-edge technology. From robotics and autonomous machines to intelligent data pipelines and predictive sensory systems, we create innovative solutions that drive long-term success for our clients. </p><p>Every woman at Emblica, including those who will join Emblica in the future, is a member of EmbliCats, bringing their unique insights and experiences to the table. Whether it&#x2019;s over breakfast meetups, coffee breaks, workshops, or other activities, we come together to exchange ideas, deepen our knowledge, support one another, and drive meaningful initiatives that empower women in tech.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2025/03/sohva2-1.jpg" class="kg-image" alt="Introducing EmbliCats: Empowering Women in Tech at Emblica" loading="lazy" width="2000" height="1335" srcset="https://blog.emblica.fi/content/images/size/w600/2025/03/sohva2-1.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2025/03/sohva2-1.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2025/03/sohva2-1.jpg 1600w, https://blog.emblica.fi/content/images/size/w2400/2025/03/sohva2-1.jpg 2400w" sizes="(min-width: 720px) 720px"><figcaption><span style="white-space: pre-wrap;">EmbliCats starting from the left: Sanna, Lotta, Venla and Anni</span></figcaption></figure><p><strong>Get to know the founding members</strong></p><p><strong>Venla</strong> &#x2013; Data Scientist and Project Manager at Emblica, delivering AI solutions to our clients. At home, I do my best to manage personal projects, such as parenting a teenager and building my camper van.</p><p><strong>Anni</strong> &#x2013; ML Engineer and Project Manager at Emblica, crafting code by day, spinning clay and bike by night. Proud cat mom of two and a certified member of EmbliCats.</p><p><strong>Lotta</strong> &#x2013; As CCO at Emblica, I help our customers make the most of AI while growing our business. Although I have lived and worked abroad and love exploring new places, my roots are on a small farm in Eastern Finland and I still relax best in nature with my fluffy dog.&#xA0;</p><p><strong>Sanna</strong> &#x2013; HR Manager at Emblica, the one who keeps things running smoothly and helps every Emblican to reach their full potential. Child of Lapland with a sunny soul shaped by Spain. You can also find me dancing, traveling, or diving into whatever new project I am currently obsessed with.</p><p><strong>What&#x2019;s Ahead</strong></p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2025/03/EmbliCats-goals.png" class="kg-image" alt="Introducing EmbliCats: Empowering Women in Tech at Emblica" loading="lazy" width="2000" height="955" srcset="https://blog.emblica.fi/content/images/size/w600/2025/03/EmbliCats-goals.png 600w, https://blog.emblica.fi/content/images/size/w1000/2025/03/EmbliCats-goals.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2025/03/EmbliCats-goals.png 1600w, https://blog.emblica.fi/content/images/size/w2400/2025/03/EmbliCats-goals.png 2400w" sizes="(min-width: 720px) 720px"></figure><p>We want to take things to the next level with an exciting lineup of events and initiatives:</p><ul><li>Collaboration &#x2013; Partnering with other female tech groups&#xA0;</li><li>Industry Engagement &#x2013; Participating in hackathons, networking events, and conferences</li><li>Knowledge Sharing &#x2013; Organizing workshops, writing blog posts, and developing internal discussions on key topics</li><li>Inspiring Future Talent &#x2013; School visits and student mentorships</li><li>Giving Back &#x2013; Charity projects that align with our mission</li><li>Fun &amp; Connection &#x2013; Hosting low-threshold events like meetups, and hangouts</li><li>Strengthening Female Community in Tech &#x2013; Creating more opportunities for women to thrive</li></ul><p>Wishing you all a wonderful International Women&#x2019;s Day!</p><p>P.S. <br>We are organizing our next open-invitation afterwork event on 03.04.2025 at 17:00, where we&#x2019;ll gather to play board games and hang out at Emblica&#x2019;s office. Read more information and sign up on <a href="https://www.linkedin.com/events/joinusforarelaxedafterworkevent7304036899520708608/" rel="noreferrer">LinkedIn</a>. Hopefully, we&#x2019;ll see you there! </p>]]></content:encoded></item><item><title><![CDATA[Data Done Right: The Foundation for Successful AI]]></title><description><![CDATA[Getting started with AI applications can be a daunting task. One big part is thinking about what data you need to achieve your goals. This blog post will help you think about how your needs inform what data you need to start integrating AI solutions into your business. ]]></description><link>https://blog.emblica.fi/data-done-right-the-foundation-for-successful-ai/</link><guid isPermaLink="false">67b89053d19eda0001cc1b6b</guid><category><![CDATA[data]]></category><category><![CDATA[AI]]></category><dc:creator><![CDATA[Rick Joosten]]></dc:creator><pubDate>Mon, 03 Mar 2025 12:26:27 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2025/02/DJI_0545_cg-Large-1.jpeg" medium="image"/><content:encoded><![CDATA[<h2 id="what-data-do-i-need"><strong>What data do I need?</strong></h2><img src="https://blog.emblica.fi/content/images/2025/02/DJI_0545_cg-Large-1.jpeg" alt="Data Done Right: The Foundation for Successful AI"><p>The best course of action is often to think about your problem before considering the data. &#x201C;I think AI should go through all of our data and identify problems&#x201D; is a sentiment we hear often but this creates an unfocused objective. In this <a href="https://blog.emblica.fi/the-you-get-what-you-order-problem-in-data-projects/"><u>blog post,</u></a> we discussed how having a clear problem definition will help you create better solutions.&#xA0;</p><p>Once the problem is clearly defined it is time to think about the data. The quality of the data used to create your algorithmic solution will directly affect the quality that the solution outputs. Shortly put: Garbage in, garbage out.&#xA0; One simple way to test whether your data is up to the task is by checking if an expert can solve the problem given the data.&#xA0;</p><p>If a human can solve a problem with the given data, so can a computer. A computer is generally faster and more efficient which is why we usually want to automate things in the first place but a human can often easily tell whether the data is garbage or not.&#xA0;&#xA0;</p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2025/02/_V1A9741.jpg" class="kg-image" alt="Data Done Right: The Foundation for Successful AI" loading="lazy" width="2000" height="1333" srcset="https://blog.emblica.fi/content/images/size/w600/2025/02/_V1A9741.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2025/02/_V1A9741.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2025/02/_V1A9741.jpg 1600w, https://blog.emblica.fi/content/images/size/w2400/2025/02/_V1A9741.jpg 2400w" sizes="(min-width: 720px) 720px"></figure><p>There however is one major area where computers outperform human experts and that is in combining information from several data sources. A senior mechanic might be able to tell how an engine is performing based on the sound alone. This comes with years of experience and is hard to teach to a new guy. So an expert can tell the performance based on sound. So what if we outfit the machine with not only a microphone but also other sensors that measure pressure, temperature, and vibrations? This additional information would allow an AI system to be much more precise in estimating the engine quality which would normally be invisible to the engineer.&#xA0;</p><p><strong>Collecting data the smart way</strong></p><p>So what if I know exactly what data I need but I don&#x2019;t already have the required data? The most straightforward way is to simply start collecting data. However, some AI techniques require a large amount of data which might be cost- or time-prohibitive to collect. Instead of manually collecting data, it might be possible to generate data using a simulated environment. This could mean simulating a full environment using programs such as Blender or Unity. For other applications, we can create statistical models where we can sample as much data as we need.&#xA0;</p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2025/02/emblica1.jpg" class="kg-image" alt="Data Done Right: The Foundation for Successful AI" loading="lazy" width="1456" height="816" srcset="https://blog.emblica.fi/content/images/size/w600/2025/02/emblica1.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2025/02/emblica1.jpg 1000w, https://blog.emblica.fi/content/images/2025/02/emblica1.jpg 1456w" sizes="(min-width: 720px) 720px"></figure><p>Using Blender you can simulate complete environments. You are in control of the light, background, reflectivity of objects, and even their exact size and shape. If your goal is to detect objects in a warehouse, if a person walks around with a camera taking pictures of objects all day it would not get the variety of data that you might need for the system to work in every environment. On top of that, you also immediately have the correct labels. Read more about this in our <a href="https://blog.emblica.fi/on-data-augmentation/"><u>blog post about data augmentation</u></a>.&#xA0;</p><p><strong>Do I need to label my data?</strong></p><p>Utilizing synthetic data immediately produces ground truth labels that can be used in supervised learning techniques. However, synthetic data cannot be used to solve all problems. In these cases, it is worth considering whether techniques that do not require labeled data can be used.&#xA0;</p><p>To illustrate, let&apos;s consider the task of detecting bad welds from pictures taken with a simple phone camera. One way is to have a group of people label a large set of pictures as good or bad. However, this can be error-prone and time-consuming. Using unsupervised learning techniques allows us to cluster the pictures based on the distinctive features of each weld without the need for labeled data. Good welds would be clustered together away from bad welds making it possible to tell good welds from bad.</p><p>Lastly, foundational models such as large language models can also be used to build solutions without having to introduce any of your own data at all. The boom in language models such as ChatGPT, Gemini, and DeepSeek has made tasks such as summarizing or sentiment analysis much more accessible. Privacy concerns aside, you could give it your text and get a good result. This gives a chance to build low-cost experiments without having to train a whole model yourself.&#xA0;&#xA0;</p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2025/02/data_labeling.png" class="kg-image" alt="Data Done Right: The Foundation for Successful AI" loading="lazy" width="1262" height="737" srcset="https://blog.emblica.fi/content/images/size/w600/2025/02/data_labeling.png 600w, https://blog.emblica.fi/content/images/size/w1000/2025/02/data_labeling.png 1000w, https://blog.emblica.fi/content/images/2025/02/data_labeling.png 1262w" sizes="(min-width: 720px) 720px"></figure><p><strong>How to get started building my AI solution</strong></p><p>When building AI solutions for your business problem the data required to make it happen should be considered. Here is my suggestion for how to utilize data when starting to build an AI application.&#xA0;</p><ol><li>Check if foundational models can be used to solve your problem. This might include using public-domain data or small sets of your own data. This is often possible in the computer vision or language domain.&#xA0;</li><li>If there is some data available specifically for your task, try building a simple model and check the quality of the model. If the model performs okay but does not match the requirements, this can be a good indicator that more quality data is required to get the results you need</li><li>After testing with either step 1 or 2 and the results look ok but not good enough it is worth exploring if synthetic data can be employed to make your model fit exactly to your problem.&#xA0;</li><li>Lastly, push for the last improvement in accuracy by labeling a small set of data by hand. This can bridge the gap between the synthetic data and real-world data.&#xA0;</li></ol><p>Want to know how this can be applied to your specific problem? Please reach out to us at Emblica and we can discuss what would fit best to your problem. </p><hr><p><a href="emblica.com"><em>Emblica</em></a><em> is not your average data team. We build customized solutions for collecting, processing, and utilizing data for all sectors, especially at the R&amp;D interface. Whether our target is a factory line, an online store or a field, you can find us busy at work, hands in the clay - at least at our office in Helsinki.</em><br></p>]]></content:encoded></item><item><title><![CDATA[Looking Back at 2024: A Year of Growth and New Adventures]]></title><description><![CDATA[2024 has been a transformative year for Emblica, marked by growth, innovation, and significant milestones. The CEO of Emblica reflects on the year’s highlights, including key events, impactful partnerships and the dedicated team behind our projects. ]]></description><link>https://blog.emblica.fi/looking-back-at-2024-a-year-of-growth-and-new-adventures/</link><guid isPermaLink="false">67643374f6e33a0001e1fbbd</guid><category><![CDATA[recap]]></category><category><![CDATA[emblica]]></category><category><![CDATA[Growth]]></category><dc:creator><![CDATA[Teemu Heikkilä]]></dc:creator><pubDate>Fri, 20 Dec 2024 12:09:47 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2024/12/sailing5-1.jpg" medium="image"/><content:encoded><![CDATA[<img src="https://blog.emblica.fi/content/images/2024/12/sailing5-1.jpg" alt="Looking Back at 2024: A Year of Growth and New Adventures"><p>As we wrap up 2024, I&#x2019;m excited to reflect on what has been a truly rewarding year for all of us at Emblica. We&#x2019;ve grown in many ways&#x2014;welcoming new faces to the team, exploring new markets, and pushing forward on projects we&#x2019;re passionate about. It&#x2019;s been a year full of hard work, learning, and plenty of good times together.</p><p>One of our big goals for 2024 was expanding into Germany, and I&#x2019;m happy to say we&#x2019;ve made great progress. We took part in events like <strong>SMM Hamburg</strong>, the leading maritime trade fair, building some fantastic connections along the way. Joining <strong>AHK</strong>, the German-Finnish Chamber of Commerce, has been another key step, leading to collaborative projects like webinars with <strong>DFKI</strong>, Germany&#x2019;s renowned AI research center, and even a keynote at the <strong>AHK Spring Forum</strong>.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2024/12/IMG_5384-Emblica.jpg" class="kg-image" alt="Looking Back at 2024: A Year of Growth and New Adventures" loading="lazy" width="2000" height="1307" srcset="https://blog.emblica.fi/content/images/size/w600/2024/12/IMG_5384-Emblica.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2024/12/IMG_5384-Emblica.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2024/12/IMG_5384-Emblica.jpg 1600w, https://blog.emblica.fi/content/images/2024/12/IMG_5384-Emblica.jpg 2000w" sizes="(min-width: 720px) 720px"><figcaption><span style="white-space: pre-wrap;">AHK Spring Forum keynote, photo by AHK / Friedrich von der Hagen</span></figcaption></figure><p>Speaking of DFKI, we launched <a href="https://blog.emblica.fi/aid-forge/" rel="noreferrer"><strong>AI&#x2019;d Forge</strong></a> together this year&#x2014;a joint program that combines our expertise in AI with their deep research capabilities. We&#x2019;ve also been able to connect with like-minded organizations by becoming a part of <strong>SIX MWMs Cluster</strong> and <strong>AI Finland.</strong></p><p>Another highlight of the year was being recognized in the <strong>Kasvu Open Growth Path Programme</strong> where we made it to the <a href="https://blog.emblica.com/inside-emblicas-kasvu-open-success-insights-from-lotta-and-teemu/" rel="noreferrer"><strong>TOP 10</strong></a>. The program gave us fresh ideas and confirmed that we&#x2019;re on the right track with how we&#x2019;re growing the company.</p><p>On the home front, our team has grown with amazing new colleagues, and expanding the team continues next year. It&#x2019;s been great to see the energy and ideas they&#x2019;ve brought to the table. </p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2024/12/_V1A9513--2-.jpg" class="kg-image" alt="Looking Back at 2024: A Year of Growth and New Adventures" loading="lazy" width="2000" height="1240" srcset="https://blog.emblica.fi/content/images/size/w600/2024/12/_V1A9513--2-.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2024/12/_V1A9513--2-.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2024/12/_V1A9513--2-.jpg 1600w, https://blog.emblica.fi/content/images/2024/12/_V1A9513--2-.jpg 2102w" sizes="(min-width: 720px) 720px"></figure><p>We&#x2019;ve also kept our team spirit strong with a variety of events&#x2014;from a <strong>team-building trip to Tallinn</strong> and a <strong>remote work week at Emblim&#xF6;kki</strong> to sailing, painting, and even a Halloween party. I also want to give a shoutout to <strong>Emblica Ladies</strong>, who started gathering this year to discuss ways to improve opportunities for women in tech and make Emblica an even better place to work.</p><p>All in all, 2024 has been a year of progress, challenges, and plenty of reasons to celebrate. A huge thank you to our clients, partners, and collaborators&#x2014;you&#x2019;ve played an important part in everything we&#x2019;ve achieved. Here&#x2019;s to an exciting year ahead, filled with new projects, new connections, and more moments worth celebrating.</p><p>Wishing you all the best for the year to come,<br><br>Teemu Heikkil&#xE4;<br>CEO of Emblica</p><hr><p><a href="emblica.com"><em>Emblica</em></a><em> is not your average data team. We build customized solutions for collecting, processing, and utilizing data for all sectors, especially at the R&amp;D interface. Whether our target is a factory line, an online store or a field, you can find us busy at work, hands in the clay - at least at our office in Helsinki.</em></p>]]></content:encoded></item><item><title><![CDATA[Inside Emblica’s Kasvu Open success, insights from Lotta and Teemu]]></title><description><![CDATA[Emblica participated recently in Kasvu Open, a growth-focused program designed to help companies enhance their strategies and scale their operations. Lotta and Teemu share their experiences, insights gained, and what reaching the top 10 means for Emblica.]]></description><link>https://blog.emblica.fi/inside-emblicas-kasvu-open-success-insights-from-lotta-and-teemu/</link><guid isPermaLink="false">6736f4d88cdf6900018a7df5</guid><category><![CDATA[Kasvu Open]]></category><category><![CDATA[Growth]]></category><category><![CDATA[Strategy]]></category><dc:creator><![CDATA[Rick Joosten]]></dc:creator><pubDate>Fri, 15 Nov 2024 12:45:54 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2024/11/240828-master-HKI-122-1.jpg" medium="image"/><content:encoded><![CDATA[<img src="https://blog.emblica.fi/content/images/2024/11/240828-master-HKI-122-1.jpg" alt="Inside Emblica&#x2019;s Kasvu Open success, insights from Lotta and Teemu"><p>In this blog post, we sit down with Lotta (CCO) and Teemu (CEO) to discuss Emblica&#x2019;s recent participation in Kasvu Open, a growth-focused program designed to help companies enhance their strategies and scale their operations. Lotta and Teemu share their experiences, insights gained, and what reaching the top 10 means for Emblica.</p><p><strong>Kasvu Open describes itself as a program that &#x201C;organizes sparring sessions for companies to enhance their growth capabilities.&#x201D; How would you describe what Kasvu Open is?</strong></p><p><strong>Lotta:</strong> Kasvu Open is a fantastic opportunity for growth-oriented companies looking to refine their strategies and gain fresh perspectives. The program connects participants with industry experts and other companies, providing immersive sparring sessions and networking opportunities. It&apos;s a platform where businesses can receive valuable insights, all aimed at enhancing their growth capabilities.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2024/11/241010-KO-TOP60-339-3.jpg" class="kg-image" alt="Inside Emblica&#x2019;s Kasvu Open success, insights from Lotta and Teemu" loading="lazy" width="2000" height="1334" srcset="https://blog.emblica.fi/content/images/size/w600/2024/11/241010-KO-TOP60-339-3.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2024/11/241010-KO-TOP60-339-3.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2024/11/241010-KO-TOP60-339-3.jpg 1600w, https://blog.emblica.fi/content/images/2024/11/241010-KO-TOP60-339-3.jpg 2200w" sizes="(min-width: 720px) 720px"><figcaption><span style="white-space: pre-wrap;">Sparring Session, photo by Kasvu Open / Matias Ulfves</span></figcaption></figure><p><strong>Can you describe in more detail what participation in Kasvu Open entails?</strong></p><p><strong>Teemu:</strong> The program spanned about six months, during which we completed online tasks and detailed questionnaires related to our growth analysis. However, the core of the experience revolved around onsite days. These included 1:1 sparring sessions with experts, group discussions, and, of course, a lot of networking during and after the sessions.</p><p><strong>Lotta:</strong> We had the flexibility to influence the topics of our sparring discussions and could choose experts based on the areas we wanted to focus on. The sparring sessions were engaging and supportive, offering concrete tips and advice while also challenging our existing strategies. It was a great mix of friendly discussions and constructive critique.</p><p><strong>Why did Emblica decide to participate in Kasvu Open this year, and what was the application process like?</strong></p><p><strong>Teemu:</strong> This year felt like the perfect time to join Kasvu Open. Emblica is growing steadily, and we have ambitious projections for the coming years, especially as we consider international expansion. We were eager to gain external insights into our strategy and to network with experts and other like-minded companies.</p><p><strong>Lotta:</strong> The application process started with filling out a comprehensive Growth Analysis (Kasvuyritysanalyysi). This analysis covers everything from basic company information to deeper insights into our team, market potential, references, and growth potential. Experts evaluated our responses and our performance during the 1:1 meetings and we were selected to the top 60 companies. Eventually, the judges selected us for the top 10. </p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2024/11/241031-Karnevaali-314-2.jpg" class="kg-image" alt="Inside Emblica&#x2019;s Kasvu Open success, insights from Lotta and Teemu" loading="lazy" width="2000" height="1333" srcset="https://blog.emblica.fi/content/images/size/w600/2024/11/241031-Karnevaali-314-2.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2024/11/241031-Karnevaali-314-2.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2024/11/241031-Karnevaali-314-2.jpg 1600w, https://blog.emblica.fi/content/images/size/w2400/2024/11/241031-Karnevaali-314-2.jpg 2400w" sizes="(min-width: 720px) 720px"><figcaption><span style="white-space: pre-wrap;">Kasvu Open Carnival October 31, 2024: Final Pitch, photo by Kasvu Open / Matias Ulfves</span></figcaption></figure><p><strong>What do you think made Emblica a good candidate for Kasvu Open to be selected into the top 10?</strong></p><p><strong>Teemu:</strong> The judges highlighted our consistent growth, clear vision, strong references, and our modern approach to leadership. I think a modern culture where people are trusted and taken care of is something that the judges would like to see more widely adopted in other organizations.</p><p><strong>Lotta:</strong> Additionally, Emblica operates in the rapidly growing data and AI market, which gives us a natural edge in scaling our business. The fact that we have a track record of 70+ implemented projects across multiple industries helps us leverage the market growth even better.</p><p><strong>What were the biggest insights you gained from the program that you are willing to share?</strong></p><p><strong>Teemu:</strong> The most interesting observation came from a mentor who said: &#x201C;It sounds like your customers are all forerunners.&#x201D; This was something we hadn&#x2019;t realized before, but it perfectly captures the type of clients we attract.</p><p><strong>Lotta:</strong> We got so many useful insights, from strategies to scale our sales team efficiently, and enhancing our brand awareness, to hearing past experiences on expanding into new markets. The direction we&#x2019;re headed looks promising and now we have even more tools in our belt to choose from. We can use these tools when there is a need to improve any of the areas as well as many professionals in our network to consult if needed.</p><p><strong>Were there any personal highlights or memorable moments you&#x2019;d like to share?</strong></p><p><strong>Lotta:</strong> One of the most rewarding aspects was when the mentors and judges asked challenging questions that I hadn&#x2019;t considered before. It pushed me to explain our current approaches and evaluate if it is really the best approach. I enjoyed seeing how all the companies went through this and how open-minded and curious Kasvu Open participants were.</p><p><strong>Teemu:</strong> My main highlight is getting to know so many growth-oriented companies, many of which are looking to grow not only in Finland but also internationally just like us.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2024/11/241031-Karnevaali-318-1.jpg" class="kg-image" alt="Inside Emblica&#x2019;s Kasvu Open success, insights from Lotta and Teemu" loading="lazy" width="2000" height="1333" srcset="https://blog.emblica.fi/content/images/size/w600/2024/11/241031-Karnevaali-318-1.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2024/11/241031-Karnevaali-318-1.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2024/11/241031-Karnevaali-318-1.jpg 1600w, https://blog.emblica.fi/content/images/size/w2400/2024/11/241031-Karnevaali-318-1.jpg 2400w" sizes="(min-width: 720px) 720px"><figcaption><span style="white-space: pre-wrap;">Kasvu Open Carnival October 31, 2024: Judges&apos; Roast, photo by Kasvu Open / Matias Ulfves</span></figcaption></figure><p><strong>Now that Kasvu Open has concluded, can we expect any significant changes in how Emblica will operate?</strong></p><p><strong>Lotta:</strong> I wouldn&#x2019;t say there will be massive changes from a customer perspective. Our team will continue designing and developing top-tier solutions with a focus on delivering tangible business value. However, we&#x2019;re constantly improving our operations, and we will definitely incorporate the key learnings from Kasvu Open to further enhance our operations.</p><p><strong>Teemu:</strong> Agreed, no drastic changes are planned. One of the most important insights from Kasvu Open is to have the confirmation that we are doing the right things. It doesn&#x2019;t mean we&#x2019;ll stagnate, though. We are continuing to evolve in ways that matter for us and for our customers.</p><p><strong>Thank you for the interview any final thoughts?</strong></p><p><strong>Lotta:</strong> The Kasvu Open experience has been an amazing journey for Emblica, helping define our growth strategy and enhancing our network of industry experts. Making it to the top 10 was incredible and shows that our vision and expertise are set to take us to the next level.</p><hr><p><a href="emblica.com"><em>Emblica</em></a><em> is not your average data team. We build customized solutions for collecting, processing, and utilizing data for all sectors, especially at the R&amp;D interface. Whether our target is a factory line, an online store or a field, you can find us busy at work, hands in the clay - at least at our office in Helsinki.</em></p>]]></content:encoded></item><item><title><![CDATA[Solving business problems with AI-enhanced LiDAR data]]></title><description><![CDATA[The use of LiDAR data can create many opportunities to solve problems requiring spatial information. This post will explore and give examples of how these techniques can be applied to solve real business challenges. ]]></description><link>https://blog.emblica.fi/solving-business-problems-with-ai-enhanced-lidar-data/</link><guid isPermaLink="false">66fcec145ddeb200017cd720</guid><category><![CDATA[AI]]></category><category><![CDATA[lidar]]></category><category><![CDATA[point clouds]]></category><category><![CDATA[forestry]]></category><category><![CDATA[agriculture]]></category><category><![CDATA[urban planning]]></category><category><![CDATA[warehouse]]></category><category><![CDATA[factory]]></category><dc:creator><![CDATA[Markku Leppälä]]></dc:creator><pubDate>Tue, 15 Oct 2024 10:09:23 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2024/10/Screenshot-2024-06-28-at-9.53.05.png" medium="image"/><content:encoded><![CDATA[<img src="https://blog.emblica.fi/content/images/2024/10/Screenshot-2024-06-28-at-9.53.05.png" alt="Solving business problems with AI-enhanced LiDAR data"><p>The use of LiDAR data can create many opportunities to solve problems requiring spatial information. <a href="https://blog.emblica.fi/points-about-point-clouds/" rel="noreferrer">In an earlier blog post</a> we dug deeper into how machine learning models can analyze 3D point clouds for several tasks such as classification, object localization, and object segmentation. This post will explore and give examples of how these techniques can be applied to solve real business challenges. </p><h2 id="a-quick-primer-on-lidar-sensors"><strong>A quick primer on LiDAR sensors</strong></h2><p>LiDAR, or Light Detection and Ranging, is a remote sensing method that uses light in the form of a pulsed laser to measure distances. This technology generates precise, three-dimensional information about the shape of the ground and its surface characteristics. The output, known as a point cloud, consists of a large number of points that represent the surface coordinates, appearing as a dense cluster of dots that collectively depict the depth and form of the physical environment.</p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2024/10/forest2.png" class="kg-image" alt="Solving business problems with AI-enhanced LiDAR data" loading="lazy" width="2000" height="1128" srcset="https://blog.emblica.fi/content/images/size/w600/2024/10/forest2.png 600w, https://blog.emblica.fi/content/images/size/w1000/2024/10/forest2.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2024/10/forest2.png 1600w, https://blog.emblica.fi/content/images/size/w2400/2024/10/forest2.png 2400w" sizes="(min-width: 720px) 720px"></figure><p>There&apos;s no doubt about the superiority of LiDARs for obtaining high-quality spatial data. Especially now that the price of this technology is affordable for various use cases. However, the systems integrating LiDARs might face an issue of overflow of unlabeled data to be processed, a problem solvable by processing the data with AI.</p><h2 id="mapping-the-cityscape"><strong>Mapping the cityscape</strong></h2><p>LiDARs produce thousands of points per second, but not all of these are relevant to the application used. Imagine taking a look at a city from a bird&#x2019;s view with an object of counting all cars. The majority of the objects seen, such as buildings, parks, and pedestrians, might not be anyhow relevant to the task. The same applies to LiDAR data, where in many cases only a minority of the data is important for the given task.</p><p>AI-enhanced systems can automate the majority, if not all, of the object recognition and movement tracking tasks, even in real-time and in complete darkness. AI systems can also be taught to filter unwanted elements in the point cloud such as noise from rain or snow. This allows the LiDAR systems to gather data no matter the weather.&#xA0;</p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2024/10/City.png" class="kg-image" alt="Solving business problems with AI-enhanced LiDAR data" loading="lazy" width="1472" height="812" srcset="https://blog.emblica.fi/content/images/size/w600/2024/10/City.png 600w, https://blog.emblica.fi/content/images/size/w1000/2024/10/City.png 1000w, https://blog.emblica.fi/content/images/2024/10/City.png 1472w" sizes="(min-width: 720px) 720px"></figure><p>Depending on the training of the AI-enhanced systems, these can offer highly flexible or very strict recognition. Coming back to the example of the city view, the AI could recognize all cars, trucks, and buses, or just convertible cars if necessary. This level of specificity is particularly useful in industries where precision is critical, such as logistics and urban traffic management.</p><p>Access to high-quality spatial data with the possibility to strip certain features in real time opens up new business opportunities. By focusing only on relevant data, companies can optimize resources and improve service delivery.</p><h2 id="more-examples"><strong>More examples</strong></h2><p>Different industries would benefit from AI-enhanced LiDAR systems. Here are some examples of how LiDAR data can be used in different cases.</p><p><strong>Forestry and Agriculture</strong>: Point clouds can be leveraged for biomass estimation. Suitable LiDARs are accurate enough to estimate the growth of individual crops. This precise measurement allows for better management of forests, planning of harvests, and monitoring of ecosystem health. Additionally, in agriculture, LiDAR can help map field topography, enabling precise irrigation planning and terrain analysis to maximize crop yield. Furthermore, having an accurate elevation map can help improve route planning through a forest and monitor ground erosion.&#xA0;</p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2024/10/forest.png" class="kg-image" alt="Solving business problems with AI-enhanced LiDAR data" loading="lazy" width="2000" height="1128" srcset="https://blog.emblica.fi/content/images/size/w600/2024/10/forest.png 600w, https://blog.emblica.fi/content/images/size/w1000/2024/10/forest.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2024/10/forest.png 1600w, https://blog.emblica.fi/content/images/size/w2400/2024/10/forest.png 2400w" sizes="(min-width: 720px) 720px"></figure><p><strong>Urban Planning and Infrastructure</strong>: In urban settings, LiDAR combined with AI proves invaluable in managing and analyzing traffic patterns and human flow. By creating detailed 3D maps of urban areas, planners can anticipate traffic jams and optimize traffic flow based on real-time data. Moreover, AI-enhanced LiDAR can assess flood risks by analyzing terrain and elevation data to predict water flow paths in extreme weather, aiding in the design of more effective water management and flood defense systems.</p><p><strong>Factory monitoring: </strong>LiDAR data is inherently private compared to the use of cameras. This opens opportunities to track people&apos;s movement in a room without infringing on their privacy. This way we can track current processes and use this data to improve efficiency. Additionally, the spatial resolution is much better than a camera&#x2019;s. This can potentially be used for full automation helping with object detection and optimized routing for autonomous robots. </p><h2 id="getting-started-with-ai-and-lidar-sensors"><strong>Getting started with AI and LiDAR sensors</strong></h2><p>Exploring how LiDAR technology works with AI shows us its big impact across different areas like forestry, city planning, and managing infrastructure. LiDAR, enhanced by AI, helps businesses handle large amounts of spatial data quickly, focusing on the important bits to make smarter decisions. For example, it can help manage forests better or improve how traffic flows in cities, making sure the data used is accurate and specifically suited for the task.</p><p>Looking forward, the combination of LiDAR data and AI techniques is promising. It could not only improve current ways of doing things but also completely change industries. Wondering if LiDAR technology could help you and your business? Get in touch and we can help you explore the possibilities.</p><hr><p><a href="https://www.emblica.com" rel="noreferrer"><em>Emblica</em></a><em> is not your average data team. We build customized solutions for collecting, processing, and utilizing data for all sectors, especially at the R&amp;D interface. Whether our target is a factory line, an online store, or a field, you can find us busy at work, hands in the clay - at least at our office in Helsinki.</em></p>]]></content:encoded></item><item><title><![CDATA[The “You Get What You Pay For” Problem in Data Projects]]></title><description><![CDATA[When faced with a textbook example of a problem solvable by algorithms, data professionals often rush to apply their favorite algorithm without questioning the initial problem formulation. The old adage rings true: “If your only tool is a hammer, every problem looks like a nail.”]]></description><link>https://blog.emblica.fi/the-you-get-what-you-order-problem-in-data-projects/</link><guid isPermaLink="false">664b5130f9b7310001bc3278</guid><category><![CDATA[emblica]]></category><category><![CDATA[Churn Prediction]]></category><category><![CDATA[machine learning]]></category><category><![CDATA[ML]]></category><category><![CDATA[AI]]></category><category><![CDATA[ENG]]></category><dc:creator><![CDATA[Markku Leppälä]]></dc:creator><pubDate>Mon, 17 Jun 2024 14:47:51 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2024/06/ONGELMAN-MAARITTELY-JA-KONEOPPIMINEN-nuoli-09-2.png" medium="image"/><content:encoded><![CDATA[<img src="https://blog.emblica.fi/content/images/2024/06/ONGELMAN-MAARITTELY-JA-KONEOPPIMINEN-nuoli-09-2.png" alt="The &#x201C;You Get What You Pay For&#x201D; Problem in Data Projects"><p>When faced with a textbook example of a problem solvable by algorithms, data professionals often rush to apply their favorite algorithm without questioning the initial problem formulation. The old adage rings true: &#x201C;If your only tool is a hammer, every problem looks like a nail.&#x201D;</p><p><em>The same article has been previously published in </em><a href="https://blog.emblica.fi/saat-mita-tilaat-ongelma-dataprojekteissa/" rel="noreferrer"><em>finnish</em></a></p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2024/05/ONGELMAN-MAARITTELY-JA-KONEOPPIMINEN-HERO-UUSI-11.png" class="kg-image" alt="The &#x201C;You Get What You Pay For&#x201D; Problem in Data Projects" loading="lazy" width="2000" height="1125" srcset="https://blog.emblica.fi/content/images/size/w600/2024/05/ONGELMAN-MAARITTELY-JA-KONEOPPIMINEN-HERO-UUSI-11.png 600w, https://blog.emblica.fi/content/images/size/w1000/2024/05/ONGELMAN-MAARITTELY-JA-KONEOPPIMINEN-HERO-UUSI-11.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2024/05/ONGELMAN-MAARITTELY-JA-KONEOPPIMINEN-HERO-UUSI-11.png 1600w, https://blog.emblica.fi/content/images/size/w2400/2024/05/ONGELMAN-MAARITTELY-JA-KONEOPPIMINEN-HERO-UUSI-11.png 2400w" sizes="(min-width: 720px) 720px"></figure><p>When businesses start to use data in decision-making and optimization, defining the problem is one of the most error-prone parts of the development process. The problem and data determine the built solution and its operation. Hence, data-utilizing projects can be impacted by what I call the <strong>&#x201C;you get what you pay for&#x201D; phenomenon</strong>.</p><h2 id="how-does-the-you-get-what-you-pay-for-problem-show-itself">How does the &quot;you get what you pay for&quot; problem show itself?</h2><p></p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2024/06/01-PROBLEM-DEFINING.png" class="kg-image" alt="The &#x201C;You Get What You Pay For&#x201D; Problem in Data Projects" loading="lazy" width="2000" height="804" srcset="https://blog.emblica.fi/content/images/size/w600/2024/06/01-PROBLEM-DEFINING.png 600w, https://blog.emblica.fi/content/images/size/w1000/2024/06/01-PROBLEM-DEFINING.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2024/06/01-PROBLEM-DEFINING.png 1600w, https://blog.emblica.fi/content/images/size/w2400/2024/06/01-PROBLEM-DEFINING.png 2400w" sizes="(min-width: 720px) 720px"></figure><p>The success of data projects is directly influenced by the data utilized.</p><p>Cliched as they may be, the maxims about the impact of data quality on algorithm performance have deeply ingrained themselves in the minds of business decision-makers. In reality, &#x201C;data quality&#x201D; is a concept not fully understood by many. In all its complexity, the criteria for data quality are clearest to data professionals who understand what to demand from data and what to expect from it. These same data-savvy professionals may be blind to another stumbling block in data projects&#x2014;<strong>incorrect problem definition</strong>.</p><p>When a textbook example of an algorithm-solvable problem presents itself, even these professionals often rush to their <em>notebooks</em>. They should pause to consider <strong>whether the problem to be solved has been correctly set</strong>, even in cases that sound simple. Here the saying holds true among data professionals as well: &#x201C;If your only tool is a hammer, every problem looks like a nail.&#x201D;</p><h3 id="example-of-an-incorrectly-defined-problem">Example of an incorrectly defined problem</h3><p>Let&#x2019;s examine two example cases. Both address the same challenge: <strong> How to reduce churn.</strong></p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2024/06/02-PROBLEM-SOLVED.png" class="kg-image" alt="The &#x201C;You Get What You Pay For&#x201D; Problem in Data Projects" loading="lazy" width="2000" height="1125" srcset="https://blog.emblica.fi/content/images/size/w600/2024/06/02-PROBLEM-SOLVED.png 600w, https://blog.emblica.fi/content/images/size/w1000/2024/06/02-PROBLEM-SOLVED.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2024/06/02-PROBLEM-SOLVED.png 1600w, https://blog.emblica.fi/content/images/size/w2400/2024/06/02-PROBLEM-SOLVED.png 2400w" sizes="(min-width: 720px) 720px"><figcaption><i><em class="italic" style="white-space: pre-wrap;">Imaginary discussion between business and data science team. Churn prediction is a common project where &quot;you get what you pay for&quot; problem occurs.</em></i></figcaption></figure><p>Many companies have attempted to predict the turnover of paying customers (churn/attrition). Often, the idea is this: identify customers who are reducing or stopping their use of the service and target them with preventive actions, such as communications or offers.</p><p><strong>Forecasting customer turnover</strong> is exactly the kind of challenge that many data scientists drool over! Initially, it might sound like a useful and interesting case for data collection, but the problem formulation in the project contains a dangerous trap.</p><p>If we analyze business needs more broadly, we realize that the ultimate goal is to <strong>minimize customer turnover, not to predict it</strong>. The original problem definition did not hit the mark immediately but instead led to hastily solving the wrong issue.</p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2024/05/ONGELMAN-MAARITTELY-JA-KONEOPPIMINEN-nuoli-09.png" class="kg-image" alt="The &#x201C;You Get What You Pay For&#x201D; Problem in Data Projects" loading="lazy" width="2000" height="1125" srcset="https://blog.emblica.fi/content/images/size/w600/2024/05/ONGELMAN-MAARITTELY-JA-KONEOPPIMINEN-nuoli-09.png 600w, https://blog.emblica.fi/content/images/size/w1000/2024/05/ONGELMAN-MAARITTELY-JA-KONEOPPIMINEN-nuoli-09.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2024/05/ONGELMAN-MAARITTELY-JA-KONEOPPIMINEN-nuoli-09.png 1600w, https://blog.emblica.fi/content/images/size/w2400/2024/05/ONGELMAN-MAARITTELY-JA-KONEOPPIMINEN-nuoli-09.png 2400w" sizes="(min-width: 720px) 720px"></figure><p>But what harm does incorrect problem framing actually cause? If we start predicting customer turnover instead of minimizing it, the model built will likely learn to identify user groups at risk of leaving as intended. However, this approach also creates significant problems:</p><ol><li><strong>The model predicting churn tells us that customers are at risk of leaving but cannot directly say why.</strong><br>The model&apos;s output generalizes all departing users into the same category. Different customer groups at risk of leaving require different actions to improve customer relations, but the model in its current form does not enable distinguishing between them.<br><br>A pertinent example of incorrect risk classification is classifying the so-called sleeping customers. This group may include subscribers to a service whose subscription is running as expected, and no problems have occurred. These customers haven&#x2019;t thought about the ongoing charge or canceling the subscription. As a thought experiment, the model could interpret such passivity as a sign of dissatisfaction and incorrectly place a well-functioning customer relationship in the danger zone. In the worst case, the company takes action and launches a targeted marketing campaign that annoys the customers and leads them to cancel the entire service&#x2014;even though they had been completely satisfied until now.</li><li><strong>In some cases, the model functions exactly as it should, predicting departure with certainty&#x2014;and nothing can be done</strong>. Some customers may indeed be lost cases, regardless of whether they are targeted with actions or not. The model works, but actually, no customer relationship management does. Money is wasted on actions, but no results are produced.</li><li>When the model is designed, it is often considered which signals could classify a user as &#x201C;in the danger zone.&#x201D; Frequently, only user-created signals, such as those generated by using the service, are chosen. However, in designing models that solve this type of problem setting, actions from the service towards the user are rarely considered. When the prediction model is utilized, <strong>the analysis-based actions taken towards customers contaminate the data</strong>. <br><br>Suppose that a grocery delivery service&#x2019;s customer service representative reviews loyalty accounts identified by the model as at risk of leaving. The customer service representative notices that some at-risk accounts had delivery problems and as a corrective action he slips them all discount coupons. As a result, customers remain service users, but the model does not understand the reason behind the customers&#x2019; retention. When the model is retrained with new data, similar cases may no longer be classified in the same way, because according to the model&#x2019;s data, delivery issues did not cause the churn.</li></ol><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2024/06/03-PROBLEM-GROUPS.png" class="kg-image" alt="The &#x201C;You Get What You Pay For&#x201D; Problem in Data Projects" loading="lazy" width="1964" height="1198" srcset="https://blog.emblica.fi/content/images/size/w600/2024/06/03-PROBLEM-GROUPS.png 600w, https://blog.emblica.fi/content/images/size/w1000/2024/06/03-PROBLEM-GROUPS.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2024/06/03-PROBLEM-GROUPS.png 1600w, https://blog.emblica.fi/content/images/2024/06/03-PROBLEM-GROUPS.png 1964w" sizes="(min-width: 720px) 720px"><figcaption><i><em class="italic" style="white-space: pre-wrap;">Churn prediction model is able to identify at-risk customers but cannot tell what should be done to retain them.</em></i></figcaption></figure><h2 id="how-can-the-%E2%80%9Cyou-get-what-you-pay-for%E2%80%9D-problem-be-turned-into-an-advantage">How can the &#x201C;you get what you pay for&#x201D; problem be turned into an advantage?</h2><p></p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2024/06/04-PROBLEM-SOLVED.png" class="kg-image" alt="The &#x201C;You Get What You Pay For&#x201D; Problem in Data Projects" loading="lazy" width="2000" height="1295" srcset="https://blog.emblica.fi/content/images/size/w600/2024/06/04-PROBLEM-SOLVED.png 600w, https://blog.emblica.fi/content/images/size/w1000/2024/06/04-PROBLEM-SOLVED.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2024/06/04-PROBLEM-SOLVED.png 1600w, https://blog.emblica.fi/content/images/size/w2400/2024/06/04-PROBLEM-SOLVED.png 2400w" sizes="(min-width: 720px) 720px"></figure><h3 id="design-the-problem-chain-from-start-to-finish-and-consider-what-you-want-to-achieve">Design the problem chain from start to finish and consider what you want to achieve.</h3><p>&#x201C;You get what you pay for&#x201D; is not only a constant challenge in data projects but also a fundamental nature guiding their successful implementation. Defining the problem and objectives goes a long way.</p><blockquote>Changing the problem to be solved also changes the utilization approach. Ultimately, this may be the most challenging part of implementing a great algorithm, as people must also trust the machine&#x2019;s assessment of actions, even if it feels intuitively wrong.</blockquote><p>The problem should be defined from the beginning to depict the desired outcome, not an intermediary stage. In our example case, the right solution would be to build a model that <strong>predicts the best possible way to manage customer relationships</strong>, revealing what kind of treatment satisfied and dissatisfied customers really need. Changing the problem to be solved also changes the utilization approach. Ultimately, this could be the most challenging part of adopting the new algorithm, as people must trust the machine&#x2019;s evaluation of actions, even if it seems counterintuitive.</p><p>Concretely, a small change in problem definition significantly alters the entire task&#x2019;s outcomes. In our example case, comparing two different problem descriptions is relatively straightforward; we no longer aim to simply divide the entire customer base binary, but make personalized decisions for different customers.</p><p>Additionally, we shift the assessment of effects to be processed by algorithms rather than burdening people with it. In some situations, it may be possible to automate certain customer relationship management methods in the same pipeline, thereby further reducing the workload on people.</p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2024/06/05-PROBLEM-SOLVED.png" class="kg-image" alt="The &#x201C;You Get What You Pay For&#x201D; Problem in Data Projects" loading="lazy" width="1600" height="988" srcset="https://blog.emblica.fi/content/images/size/w600/2024/06/05-PROBLEM-SOLVED.png 600w, https://blog.emblica.fi/content/images/size/w1000/2024/06/05-PROBLEM-SOLVED.png 1000w, https://blog.emblica.fi/content/images/2024/06/05-PROBLEM-SOLVED.png 1600w" sizes="(min-width: 720px) 720px"></figure><p>One challenge in the first example was that customer relationship management actions contaminated the data, making the results unreliable. In this case, the model is specifically created for assessing customer relationship management, so it learns from successful and failed actions to become a better predictor. The model can evaluate the effects of various management approaches on the customer and identify which actions are best for each customer at different times. The model might identify that the best approach for most customers could simply be to leave them alone. The previous model could not do this and might have incorrectly presented them as being at risk of leaving.</p><blockquote>The simple solution to the &#x201C;you get what you pay for&#x201D; problem: change your order.</blockquote><h2 id="in-a-nutshell-how-do-i-%E2%80%9Cget-what-i-pay-for%E2%80%9D-and-benefit-from-it">In a nutshell: how do I &#x201C;get what I pay for&#x201D; and benefit from it?</h2><p></p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2024/06/06-PROBLEM-SETTING-SUMMARY.png" class="kg-image" alt="The &#x201C;You Get What You Pay For&#x201D; Problem in Data Projects" loading="lazy" width="2000" height="1125" srcset="https://blog.emblica.fi/content/images/size/w600/2024/06/06-PROBLEM-SETTING-SUMMARY.png 600w, https://blog.emblica.fi/content/images/size/w1000/2024/06/06-PROBLEM-SETTING-SUMMARY.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2024/06/06-PROBLEM-SETTING-SUMMARY.png 1600w, https://blog.emblica.fi/content/images/size/w2400/2024/06/06-PROBLEM-SETTING-SUMMARY.png 2400w" sizes="(min-width: 720px) 720px"></figure><ol><li><strong>PROBLEM</strong>. Spend time defining the problem and ensure you are addressing the right issue. A simple flowchart can be a good tool for defining objectives, operational areas, available data, and desired outcomes.</li><li><strong>DATA</strong>. Ensure that you are collecting data on aspects that are crucially related to customer relationships and your defined objectives. If data is not yet available, can it be collected?</li><li><strong>TEST</strong>. Initially, testing can be done &#x201C;on paper.&#x201D; A good guideline is to consider: &#x201C;Does this outcome directly help solve the original problem?&#x201D; Even &#x201C;hard&#x201D; models can be tested up to production alongside business operations without too much disruption.</li></ol><p>There is a simple solution to the &#x201C;you get what you pay for&#x201D; problem: if the project does not produce the desired result, change what you order. In our example, the issue was minimizing customer churn. However, because of the methods used, such as utilizing existing data and optimizing metrics, this problem often recurs in various algorithmic solutions. Therefore, it is crucial to recognize the risk posed by hasty problem definition and try to circumvent it.</p><hr><p><a href="emblica.com"><em>Emblica</em></a><em> is a technology company focused on data-intensive applications and artificial intelligence. Our customers are e.g. Sanoma, Uponor, Caruna, and the Tax Administration. Emblica is 100% owned by its employees.</em><br></p>]]></content:encoded></item><item><title><![CDATA[AI’d Forge – German-Finnish cooperation in the transfer of AI technologies to companies]]></title><description><![CDATA[AI’d Forge – German-Finnish cooperation in the transfer of AI technologies to companies]]></description><link>https://blog.emblica.fi/aid-forge/</link><guid isPermaLink="false">665de0fcf9b7310001bc32d5</guid><category><![CDATA[News]]></category><category><![CDATA[DE]]></category><category><![CDATA[machine learning]]></category><category><![CDATA[ENG]]></category><dc:creator><![CDATA[Emblica]]></dc:creator><pubDate>Mon, 03 Jun 2024 09:00:00 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2024/06/IMG_3284-1.jpg" medium="image"/><content:encoded><![CDATA[<figure class="kg-card kg-bookmark-card kg-card-hascaption"><a class="kg-bookmark-container" href="https://www.dfki.de/web/news/aid-forge-deutsch-finnische-zusammenarbeit"><div class="kg-bookmark-content"><div class="kg-bookmark-title">AI&#x2019;d Forge &#x2013; Deutsch-finnische Zusammenarbeit beim Transfer von KI-Technologien in Unternehmen</div><div class="kg-bookmark-description">Das Deutsche Forschungszentrum f&#xFC;r K&#xFC;nstliche Intelligenz (DFKI) und der finnische Experte f&#xFC;r KI-Implementierung Emblica mit Sitz in Helsinki werden k&#xFC;nftig Unternehmen bei der digitalen Transformation unterst&#xFC;tzen. Das DFKI bringt dabei seine Forschungsexpertise in den Technologiefeldern Datenanalyse, Generative KI, Bild- und Signalerkennung- und -verarbeitung sowie KI-basierte Prognostik ein. Emblica steuert seine Erfahrung in der Beratung von Unternehmen im Bereich der Gesch&#xE4;ftsprozessoptimierung durch KI-basierte Softwarel&#xF6;sungen und deren Implementierung bei. Die Zusammenarbeit wird erg&#xE4;nzt durch das Beratungsunternehmen Remode ME und deren Expertise in der internationalen Gesch&#xE4;ftsentwicklung und im Venture Building. Das gemeinsame Programm unter der Bezeichnung &#x201E;AI&#x2019;d Forge&#x201C; wurde am 30. Mai 2024 auf der Digital Convention &#x201C;Noerd&#x201D; in Rostock erstmals vorgestellt.</div><div class="kg-bookmark-metadata"><img class="kg-bookmark-icon" src="https://www.dfki.de/typo3conf/ext/wftmpl_dfki/Resources/Public/Images/favicon.ico" alt="AI&#x2019;d Forge &#x2013; German-Finnish cooperation in the transfer of AI technologies to companies"></div></div><div class="kg-bookmark-thumbnail"><img src="https://www.dfki.de/fileadmin/_processed_/a/2/csm_Philipp_Teemu_6c5a14813f.jpeg" alt="AI&#x2019;d Forge &#x2013; German-Finnish cooperation in the transfer of AI technologies to companies"></div></a><figcaption><img src="https://blog.emblica.fi/content/images/2024/06/IMG_3284-1.jpg" alt="AI&#x2019;d Forge &#x2013; German-Finnish cooperation in the transfer of AI technologies to companies"><p dir="ltr"><span style="white-space: pre-wrap;">Auf Deutsch (DFKI.de)</span></p></figcaption></figure><p>The German Research Center for Artificial Intelligence (DFKI) and Finnish AI implementation expert Emblica, based in Helsinki, will support companies in digital transformation in the future. DFKI will contribute its research expertise in the technology fields of data analysis, generative AI, image and signal recognition and processing as well as AI-based forecasting. Emblica will share its experience in advising companies in the field of business process optimization through AI-based software solutions and their implementation. The collaboration is complemented by the consulting company Remode ME and their expertise in international business development and venture building. The joint program under the name &quot;AI&apos;d Forge&quot; was presented for the first time May 30, 2024 at the digital convention &quot;Noerd&quot; in Rostock.</p><p>AI&apos;d Forge aims to support companies in identifying the potential of artificial intelligence in their operations, developing customized AI solutions and implementing them.</p><p>Emblica focuses, in particular, on those areas in which AI promises a high return on investment and scaling opportunities. DFKI&apos;s broad research spectrum and Emblica&apos;s many years of practical experience offer companies a combination of in-depth industry knowledge and the latest research results. The aim is to implement these in such a way that the result is seamless integration into the existing company software.</p><p>Philipp Koch, Research Department Manager in DFKI&#x2019;s AI in Medical Image and Signal Processing research area:</p><p><em>&#x201C;We aim to support companies in the best possible way with the use of artificial intelligence through AI&apos;d Forge. In doing so, we rely not only on the expertise and know-how of the DFKI but also on the experience of Emblica to jointly identify challenges and develop tailored solutions for companies. Our focus is on a trustworthy and needs-based collaboration as well as knowledge transfer so that companies can optimally leverage AI opportunities for their success.&#x201D;</em></p><p>Teemu Heikkil&#xE4;, CEO of Emblica:</p><p><em>&quot;With AI&apos;d Forge, we support companies throughout their individual journey to the optimal use of artificial intelligence - from the initial idea to the implementation of the solution, its operation and its continuous further development. We always consider a company&#x2019;s needs and resources, and customize our support accordingly, including knowledge transfer to in-house developers. Together with DFKI&apos;s expertise and its broad research spectrum, we can significantly improve the way companies approach the implementation of AI.&quot;</em></p><p>To this end, DFKI and Emblica have defined a model process that holistically maps the introduction of AI processes in companies. It describes the entire process chain from knowledge transfer to the possible uses of AI through to implementation. In joint workshops, the AI&apos;d Forge team works with the company to identify the business areas with potential for optimization. The next step is to examine which AI technologies and processes could provide a remedy. Before implementation, AI&apos;d Forge draws up a business case that specifically elaborates the expected benefits. The DFKI and Emblica team then examines what is required to develop an adapted AI solution and estimates the time, personnel and financial costs involved. The roll-out occurs based on a concrete development and implementation plan, which also focuses on the transfer of knowledge to in-house developers.</p><p>Thanks to the collaboration between a German research partner and a Finnish implementation service provider, AI&apos;d Forge offers companies from both countries access to leading expertise in the field of AI. In the future, AI&apos;d Forge could serve as a model for cross-border and cross-sector collaboration with other European partners.</p><p><strong>About DFKI</strong></p><p>The German Research Center for Artificial Intelligence GmbH (DFKI) was founded in 1988 as a non-profit public-private partnership (PPP). It has locations in Kaiserslautern, Saarbr&#xFC;cken, Bremen and Lower Saxony, laboratories in Berlin, Darmstadt and L&#xFC;beck, as well as a branch office in Trier. DFKI has been researching AI for humans for over 35 years and is oriented towards social relevance and scientific excellence in the key future-oriented research and application areas of artificial intelligence. DFKI is one of the most important &quot;Centers of Excellence&quot; in the international scientific community. Around 1,560 employees from over 76 nations are currently researching innovative software solutions. In its laboratory in L&#xFC;beck, DFKI is researching AI in medical image and signal processing.</p><p><strong>About Emblica</strong></p><p>Emblica is a Finnish company that has specialized in artificial intelligence for over ten years. It develops and implements customized AI solutions for companies in the private and public sectors. With a focus on practical implementation and measurable results, Emblica has led AI implementation in numerous market-leading organizations.</p><p><strong>About Remode ME</strong></p><p>Remode ME is a consulting company based in Helsinki that helps innovative companies in Europe and Africa achieve growth and global success. Remode ME&apos;s international team develops and implements strategies for market entry, building B2B/BB2B partnerships, and venture building.</p><p><br><strong>Further information:&#xA0;</strong><a href="https://aidforge.eu/" rel="noreferrer">https://aidforge.eu</a></p><p><br><em>(Joint press release DFKI &amp; Emblica)</em></p>]]></content:encoded></item><item><title><![CDATA[How to fool an AI system]]></title><description><![CDATA[Emblica’s AI gatekeeper for an aMazed escape room.]]></description><link>https://blog.emblica.fi/how-to-fool-an-ai-system/</link><guid isPermaLink="false">659faf23aa08900001df2824</guid><category><![CDATA[Computer Vision]]></category><category><![CDATA[Data augmentation]]></category><dc:creator><![CDATA[Rick Joosten]]></dc:creator><pubDate>Tue, 06 Feb 2024 08:40:57 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2024/01/gatekeeper_background.jpg" medium="image"/><content:encoded><![CDATA[<img src="https://blog.emblica.fi/content/images/2024/01/gatekeeper_background.jpg" alt="How to fool an AI system"><p><em>Lyhyesti suomeksi: Rakensimme </em><a href="https://www.amazed.fi/" rel="noreferrer"><em>aMazedin</em></a><em> pakopelihuoneeseen teko&#xE4;ly&#xE4; hy&#xF6;dynt&#xE4;v&#xE4;n teht&#xE4;v&#xE4;n. T&#xE4;ss&#xE4; blogissa kerromme, miten teko&#xE4;lyj&#xE4;rjestelm&#xE4;n &quot;huijattavuus&quot; liittyy esimerkiksi tehdaslinjaston laaduntarkkailuun.</em></p><p>After escaping from two of the escape rooms at <a href="https://www.amazed.fi/" rel="noreferrer">aMazed</a> during an after-work we started thinking, would it be fun to add AI elements to an escape room? We had a brief discussion about having such a feature with the team at aMazed before we even left the rooms. Shortly after, we received an email stating that aMazed was interested in working with Emblica to develop what ultimately became the AI gatekeeper. This blog will describe what the AI gatekeeper is and what we learned about fooling AI systems during its development. </p><p>Inevitably this post will contain some spoilers about the room. We tried to keep the spoilers to a minimum, but if you want to experience the room completely fresh you might want to read the rest of this post after you have escaped from the room. More information about the room <a href="https://www.amazed.fi/alchemiaenglish" rel="noreferrer">Alch&#x1EBD;mia</a> can be found on Amazed&#x2019;s website. <a href="https://www.amazed.fi/blog/2023/alchemia" rel="noreferrer">This post</a> by aMazed, however, is safe to read and tells about their experience with this AI project as well as what it takes to develop an escape room in general.</p><h1 id="the-ai-gatekeeper">The AI gatekeeper</h1><p>&#x26A0; <em>Spoilers for the escape room </em>Alch&#x1EBD;mia <em>are coming up, you have been warned. </em>&#x26A0;&#xA0;</p><p>Professor Happygolucky has developed the elixir of happiness that will take everyone&#x2019;s sorrows away. It is safely locked behind many puzzles and other safety measures so that it doesn&#x2019;t fall into the wrong hands. The task we have developed is one of such gatekeepers. You need to prove that you are the professor to our &#x201C;face recognition&#x201D; to progress through the room. Of course, there is only one professor so the task is to look enough like her for the system to let you continue.&#xA0;</p><figure class="kg-card kg-gallery-card kg-width-wide"><div class="kg-gallery-container"><div class="kg-gallery-row"><div class="kg-gallery-image"><img src="https://blog.emblica.fi/content/images/2024/01/853a5aaa-8f95-4dd6-aa3b-e741e5ea0992.jpeg" width="320" height="240" loading="lazy" alt="How to fool an AI system"></div><div class="kg-gallery-image"><img src="https://blog.emblica.fi/content/images/2024/02/ae80ad00-cecf-4d76-b791-641402842499--1-.jpeg" width="320" height="240" loading="lazy" alt="How to fool an AI system"></div></div></div></figure><p>The gatekeeper has multiple core elements. First, it uses a camera feed to detect whether the elements that are associated with the professor are in view using machine learning. Secondly, the results are shown to the players on a screen with a thematic user interface so that the users can monitor their progress. Lastly, the computer is connected to a magnetic lock that unlocks the next part of the escape room. All this is wrapped up so that the game administrator can, if necessary, control the game from their control center and reset the gatekeeper for new players.</p><p></p><h2 id="making-fooling-the-gatekeeper-possible">Making fooling the gatekeeper possible</h2><p>To make sure the puzzle is possible to solve we cannot make the gatekeeper too strict in how it determines whether to let you continue or not. If you solve the puzzle but the system does not let you continue it can ruin the fun. On the other hand, if it lets everyone through no matter what it is not a great puzzle either. The problem of finding a balance between these two extremes we face not only in this project but appears in almost every project that involves automatic decision making. Let&#x2019;s step out of the escape room to illustrate these problems so we don&#x2019;t spoil the puzzle fully.</p><p>Consider a quality control system on an assembly line with the primary aim of ensuring that only products with all necessary components correctly assembled are shipped to customers. Striking the right balance is crucial; if the system is too lenient, it risks sending out defective products, while being too strict may lead to the rejection of perfectly functional items, increasing costs for the factory. How we collect data to train these models plays a pivotal role in determining how the system behaves. For example, if our training data comprises images taken from the assembly line with consistent angles and lighting conditions, the system may only recognize products that look identical. However, if a product arrives slightly rotated or a light breaks in the factory, the system might struggle to identify it as a good product and reject it. On the flip side, if our data only includes extremely faulty cases without instances of minor defects, the model might miss the nuances of what constitutes a good product, potentially allowing some faulty items to slip through and be shipped to the customer.</p><figure class="kg-card kg-gallery-card kg-width-wide"><div class="kg-gallery-container"><div class="kg-gallery-row"><div class="kg-gallery-image"><img src="https://blog.emblica.fi/content/images/2021/08/blogi-augmentation-1.jpg" width="1912" height="1232" loading="lazy" alt="How to fool an AI system" srcset="https://blog.emblica.fi/content/images/size/w600/2021/08/blogi-augmentation-1.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2021/08/blogi-augmentation-1.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2021/08/blogi-augmentation-1.jpg 1600w, https://blog.emblica.fi/content/images/2021/08/blogi-augmentation-1.jpg 1912w" sizes="(min-width: 720px) 720px"></div><div class="kg-gallery-image"><img src="https://blog.emblica.fi/content/images/2021/08/blogi-augmentation-2.jpg" width="1912" height="1232" loading="lazy" alt="How to fool an AI system" srcset="https://blog.emblica.fi/content/images/size/w600/2021/08/blogi-augmentation-2.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2021/08/blogi-augmentation-2.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2021/08/blogi-augmentation-2.jpg 1600w, https://blog.emblica.fi/content/images/2021/08/blogi-augmentation-2.jpg 1912w" sizes="(min-width: 720px) 720px"></div></div></div></figure><p>Making sure the training data covers all sorts of situations it might face is crucial. One technique to create more diverse training data is data augmentation. You can dive into the nitty-gritty details in our blog post <a href="https://blog.emblica.fi/on-data-augmentation/"><u>here</u></a>. In short, instead of keeping things too predictable, we throw in different angles, lighting scenarios, and maybe even simulate some hiccups. It&apos;s like giving our AI system a taste of real-life unpredictability, whether it&apos;s dealing with an escape room or managing an assembly line.</p><figure class="kg-card kg-gallery-card kg-width-wide"><div class="kg-gallery-container"><div class="kg-gallery-row"><div class="kg-gallery-image"><img src="https://blog.emblica.fi/content/images/2021/08/blogi-augmentation-3.jpg" width="1912" height="1232" loading="lazy" alt="How to fool an AI system" srcset="https://blog.emblica.fi/content/images/size/w600/2021/08/blogi-augmentation-3.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2021/08/blogi-augmentation-3.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2021/08/blogi-augmentation-3.jpg 1600w, https://blog.emblica.fi/content/images/2021/08/blogi-augmentation-3.jpg 1912w" sizes="(min-width: 720px) 720px"></div><div class="kg-gallery-image"><img src="https://blog.emblica.fi/content/images/2021/08/blogi-augmentation-4.jpg" width="1912" height="1232" loading="lazy" alt="How to fool an AI system" srcset="https://blog.emblica.fi/content/images/size/w600/2021/08/blogi-augmentation-4.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2021/08/blogi-augmentation-4.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2021/08/blogi-augmentation-4.jpg 1600w, https://blog.emblica.fi/content/images/2021/08/blogi-augmentation-4.jpg 1912w" sizes="(min-width: 720px) 720px"></div></div></div></figure><p>Now, let&#x2019;s get back to the escape room. To make sure that our model makes a good puzzle we collected a diverse set of training data together with people from aMazed, and used different augmentation techniques. Is the puzzle hard enough, but still fun to solve? You will have to find it out by yourselves and see if you can fool the AI gatekeeper!</p><hr><p><a href="emblica.com"><em>Emblica</em></a><em> is a technology company focused on data-intensive applications and artificial intelligence. Our customers are e.g. Sanoma, Uponor, Caruna, and the Tax Administration. Emblica is 100% owned by its employees.</em></p>]]></content:encoded></item><item><title><![CDATA[Remote-remote working]]></title><description><![CDATA[“We support remote work” means very different things in different organizations. After cycling from Finland to Morocco for 5 months, Vili is interviewed by Anna about his experience of working full-time while changing locations often.]]></description><link>https://blog.emblica.fi/remote-remote-working/</link><guid isPermaLink="false">64429188cb2c1c0001838c3e</guid><category><![CDATA[remote]]></category><category><![CDATA[hybrid-work]]></category><category><![CDATA[company culture]]></category><category><![CDATA[ENG]]></category><dc:creator><![CDATA[Vili Hätönen]]></dc:creator><pubDate>Fri, 21 Apr 2023 13:41:00 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2023/04/Screenshot-from-2023-04-21-18-41-31.png" medium="image"/><content:encoded><![CDATA[<img src="https://blog.emblica.fi/content/images/2023/04/Screenshot-from-2023-04-21-18-41-31.png" alt="Remote-remote working"><p><em>Lyhyesti suomeksi: Kun yritys sanoo &#x201C;tukevansa et&#xE4;ty&#xF6;skentely&#xE4;&#x201D;, se voi tarkoittaa hyvin eri asioita. T&#xE4;ss&#xE4; blogissa Anna haastattelee Suomesta Marokkoon py&#xF6;r&#xE4;illytt&#xE4; varatoimistusjohtajaamme Vili&#xE4;.</em></p><p>Over the past three years, remote working has been discussed a lot. For an organization like Emblica, remote working was nothing new when the Covid-19 pandemic hit Finland in early 2020. Once the pandemic had forced many organizations to take a leap in their digitalization and remote work practices, consulting roles are now widely expected to allow hybrid (remote + on-site) work. However, the reality of what hybrid work looks like varies a lot. </p><p>To push the limits of remote work, and as a personal experiment, Emblica&apos;s vice CEO Vili cycled from North Europe to Africa while working full-time. Our People Happiness Officer Anna interviewed Vili regarding his personal experiences of this remote-remote work experiment. </p><p><strong>Hi Vili! Thanks for joining me today to discuss remote work experiences at Emblica. You had a little cycling trip last year, could you tell us about it? </strong></p><p>Yes, we decided to try an unusual remote-remote working setup last year. By remote-remote working, I mean someone working remotely from somewhere else than their default remote working location, which would be usually their home. I had an idea to cycle to Africa while simultaneously working full-time, and we decided to give it a go. </p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2023/04/image.png" class="kg-image" alt="Remote-remote working" loading="lazy" width="649" height="1092" srcset="https://blog.emblica.fi/content/images/size/w600/2023/04/image.png 600w, https://blog.emblica.fi/content/images/2023/04/image.png 649w"><figcaption><span style="white-space: pre-wrap;">A social media update, the picture is from the coast of Denmark.</span></figcaption></figure><p><strong>How did you come up with the idea to cycle to Africa? How far away is that from Helsinki?</strong></p><p>Well, the previous year, someone asked me how I would get to my volunteering post in Namibia, and asked if I would cycle there. I thought they were crazy, and I told them that. But the idea had been planted, so a year or so later, I decided to act on it and cycle the 5000km from Finland to North Africa via ten different countries. Which probably makes me the crazy one. </p><p><strong>Yes, it sort of does. To clarify, we don&apos;t push or expect our employees to be like this. Also, we want to ensure everyone&apos;s well-being is taken care of from all aspects, also while remote working. So how did you plan and prepare for the trip in advance?</strong></p><p>The biggest part of the planning was organizing projects so that the trip would not bother customers and our internal team dynamics. It took me half a year to sort these out. This was really helpful since it motivated me to document certain things better and share responsibilities. The remote-remote working equipment was light: a laptop, a charger, a hard drive, and some clothes appropriate for visiting an office somewhere along the way. You don&apos;t need more than that, although you need to look after your work ergonomics when working in different locations with often non-optimal tables and chairs.</p><p>I didn&apos;t really prepare for the cycling, but I figured that I will learn along the way. I originally estimated the duration and route by checking a map app for a route from Finland to South Spain and dividing that by 100km, which was my guess for a daily cycling distance. Then add the work days between the cycling weekends and <em>tada</em>, I had my estimate of 5 months on the road.</p><figure class="kg-card kg-video-card kg-width-regular kg-card-hascaption" data-kg-thumbnail="https://blog.emblica.fi/content/images/2023/04/media-thumbnail-ember1424.jpg" data-kg-custom-thumbnail>
            <div class="kg-video-container">
                <video src="https://blog.emblica.fi/content/media/2023/04/hateless_bikemore_FIN_MOR_2022_1to1.mp4" poster="https://img.spacergif.org/v1/1274x1274/0a/spacer.png" width="1274" height="1274" loop autoplay muted playsinline preload="metadata" style="background: transparent url(&apos;https://blog.emblica.fi/content/images/2023/04/media-thumbnail-ember1424.jpg&apos;) 50% 50% / cover no-repeat;"></video>
                <div class="kg-video-overlay">
                    <button class="kg-video-large-play-icon" aria-label="Play video">
                        <svg xmlns="http://www.w3.org/2000/svg" viewbox="0 0 24 24">
                            <path d="M23.14 10.608 2.253.164A1.559 1.559 0 0 0 0 1.557v20.887a1.558 1.558 0 0 0 2.253 1.392L23.14 13.393a1.557 1.557 0 0 0 0-2.785Z"/>
                        </svg>
                    </button>
                </div>
                <div class="kg-video-player-container kg-video-hide">
                    <div class="kg-video-player">
                        <button class="kg-video-play-icon" aria-label="Play video">
                            <svg xmlns="http://www.w3.org/2000/svg" viewbox="0 0 24 24">
                                <path d="M23.14 10.608 2.253.164A1.559 1.559 0 0 0 0 1.557v20.887a1.558 1.558 0 0 0 2.253 1.392L23.14 13.393a1.557 1.557 0 0 0 0-2.785Z"/>
                            </svg>
                        </button>
                        <button class="kg-video-pause-icon kg-video-hide" aria-label="Pause video">
                            <svg xmlns="http://www.w3.org/2000/svg" viewbox="0 0 24 24">
                                <rect x="3" y="1" width="7" height="22" rx="1.5" ry="1.5"/>
                                <rect x="14" y="1" width="7" height="22" rx="1.5" ry="1.5"/>
                            </svg>
                        </button>
                        <span class="kg-video-current-time">0:00</span>
                        <div class="kg-video-time">
                            /<span class="kg-video-duration">0:25</span>
                        </div>
                        <input type="range" class="kg-video-seek-slider" max="100" value="0">
                        <button class="kg-video-playback-rate" aria-label="Adjust playback speed">1&#xD7;</button>
                        <button class="kg-video-unmute-icon" aria-label="Unmute">
                            <svg xmlns="http://www.w3.org/2000/svg" viewbox="0 0 24 24">
                                <path d="M15.189 2.021a9.728 9.728 0 0 0-7.924 4.85.249.249 0 0 1-.221.133H5.25a3 3 0 0 0-3 3v2a3 3 0 0 0 3 3h1.794a.249.249 0 0 1 .221.133 9.73 9.73 0 0 0 7.924 4.85h.06a1 1 0 0 0 1-1V3.02a1 1 0 0 0-1.06-.998Z"/>
                            </svg>
                        </button>
                        <button class="kg-video-mute-icon kg-video-hide" aria-label="Mute">
                            <svg xmlns="http://www.w3.org/2000/svg" viewbox="0 0 24 24">
                                <path d="M16.177 4.3a.248.248 0 0 0 .073-.176v-1.1a1 1 0 0 0-1.061-1 9.728 9.728 0 0 0-7.924 4.85.249.249 0 0 1-.221.133H5.25a3 3 0 0 0-3 3v2a3 3 0 0 0 3 3h.114a.251.251 0 0 0 .177-.073ZM23.707 1.706A1 1 0 0 0 22.293.292l-22 22a1 1 0 0 0 0 1.414l.009.009a1 1 0 0 0 1.405-.009l6.63-6.631A.251.251 0 0 1 8.515 17a.245.245 0 0 1 .177.075 10.081 10.081 0 0 0 6.5 2.92 1 1 0 0 0 1.061-1V9.266a.247.247 0 0 1 .073-.176Z"/>
                            </svg>
                        </button>
                        <input type="range" class="kg-video-volume-slider" max="100" value="100">
                    </div>
                </div>
            </div>
            <figcaption><p><span style="white-space: pre-wrap;">The 5000km route from Helsinki to Tarifa. The color indicates altitude. The bar plot shows the distance covered over time.</span></p></figcaption>
        </figure><p><strong>How long did it take you in the end?</strong></p><p>Exactly 5 months, by the day. But only because I needed to make it, not because it was an easy pace. </p><p><strong>Well, you reap what you sow. But five months is a long time. How did not seeing your colleagues affect your experience and your work? </strong></p><p>I did miss the team for sure. You had so many great activities and fun times without me! But pair coding and working together on the same topic were not affected almost at all. Hybrid work is such an ingrained part of our everyday routine. However, brainstorming and throwing ideas around spontaneously did decrease a lot. When you meet remotely, you talk business and stay on topic. There is much less room to share random thoughts about recent advances in ML when you&apos;re not physically in the same space sharing the short empty moments between work tasks. </p><p><strong>So working remotely hurts innovation?</strong></p><p>On the one hand, yes, you tend to share and brainstorm less. But on the other hand, when you are far away, you can see the big picture better. Just make sure you have time for big thoughts, the ideas that really create change. Sometimes I didn&apos;t manage to have that time, but it had a lot to do with the practical issue of me trying to survive across the continent. </p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2023/04/20221003_130143.jpg" class="kg-image" alt="Remote-remote working" loading="lazy" width="2000" height="2667" srcset="https://blog.emblica.fi/content/images/size/w600/2023/04/20221003_130143.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2023/04/20221003_130143.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2023/04/20221003_130143.jpg 1600w, https://blog.emblica.fi/content/images/size/w2400/2023/04/20221003_130143.jpg 2400w" sizes="(min-width: 720px) 720px"><figcaption><span style="white-space: pre-wrap;">Sometimes one can find enthusiastic pair coders from unexpected places. Bordeaux, France</span></figcaption></figure><p><strong>Were there any other practical things that made things difficult regarding work?</strong></p><p>Working remotely went surprisingly smoothly. Of course, you need basic things like electricity, an internet connection, and food to not lose time. Because I was crossing Europe, the electricity for charging my devices and 4G for a decent internet connection was guaranteed in the cities over 98% of the time. Honestly, organizing a lunch without spending half a day moving around the new city proved the most difficult thing.</p><p><strong>Well, as you know, sometimes organizing a team lunch can also be challenging when working in the office! Great to hear things went smoothly. How about the data privacy and security topics, how did you cover this part? </strong></p><p>The most important decision, during your travel, regarding privacy and data security is where you work. Some meetings you can have in a caf&#xE9; for sure, but most of the client work and anything HR-related needs to happen behind closed doors. In practice, that meant having a private room accessible for the whole day.</p><p><strong>Where were you staying? What kind of accommodations would you recommend for a remote worker?</strong></p><p>I was mostly staying in AirBnB&apos;s during the work week mainly for three reasons: cooking and doing laundry is so much faster when you don&apos;t need always to leave your flat; you get a calm, private space for work calls; and the wifi is not shared with a hundred other guests. Staying in a private accommodation was more efficient work-wise in many ways. </p><figure class="kg-card kg-image-card"><img src="https://blog.emblica.fi/content/images/2023/04/IMG_20221207_231734_723.jpg" class="kg-image" alt="Remote-remote working" loading="lazy" width="1080" height="1920" srcset="https://blog.emblica.fi/content/images/size/w600/2023/04/IMG_20221207_231734_723.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2023/04/IMG_20221207_231734_723.jpg 1000w, https://blog.emblica.fi/content/images/2023/04/IMG_20221207_231734_723.jpg 1080w" sizes="(min-width: 720px) 720px"></figure><p>Talking about wifi, you can never trust that it will work. Always have a 4G plan available, and your phone charged. I don&apos;t think there was a single week I didn&apos;t need to share the internet from my phone. Luckily the costs for using mobile data are minimal for Finnish plans in Europe. If you work elsewhere, be prepared to sort out your prepaid as the first thing when you arrive in a new country. </p><p><strong>Apart from internet access, what about another crucial topic: rest and work-life balance?</strong></p><p>Both are really important and, in the case of my trip, both more easy and more difficult than normal. Work-life balance was easy: when you cycle across the Danish coast or Spanish mountains, you don&apos;t think about work, and when you sit in your AirBnB, you have no distractions from work. I recovered well from work while cycling, and from cycling when I stopped for work. </p><p>But overall, my schedule was too tight to have time to recover from both. Working and cycling for five months with a handful of rest days is not really sustainable. Unless you cover a shorter distance or are in a better physical condition. On the bright side, I didn&apos;t struggle to fall asleep once during the trip.</p><p><strong>Sure, in this case, exercise was provided by the choice of transportation. </strong></p><p>I think I would have found it easy to detach from work even if I would have been traveling by train or car; being immersed in a new environment does make it easy to live in the moment and forget projects for a moment. But organizing a healthy amount of exercise would require extra effort in those cases. </p><figure class="kg-card kg-image-card kg-width-wide kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2023/04/image-2.png" class="kg-image" alt="Remote-remote working" loading="lazy" width="1691" height="745" srcset="https://blog.emblica.fi/content/images/size/w600/2023/04/image-2.png 600w, https://blog.emblica.fi/content/images/size/w1000/2023/04/image-2.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2023/04/image-2.png 1600w, https://blog.emblica.fi/content/images/2023/04/image-2.png 1691w" sizes="(min-width: 1200px) 1200px"><figcaption><span style="white-space: pre-wrap;">One gets to try many different remote working locations, and sometimes your work and travel days overlap. D&#xFC;sseldorf, Germany (left); a park bench near Valencia, Spain (middle); an old mansion in Moliets-et-Maa, France (right)</span></figcaption></figure><p><strong>Talking about extra effort, what unexpected extra effort comes with working remotely-remotely like that?</strong></p><p>Booking accommodation and planning which places to go takes surprisingly much time. And I didn&apos;t even need to book transportation. Although, having a bike did bring some maintenance and route planning burden. But regardless of the transportation, I was practically moving flats almost every week. And once you are in a new place, all everyday tasks like grocery shopping take longer. So if you&apos;re working full days, don&apos;t expect to experience the city during the week: your evenings are mostly spent staying alive and preparing for the next change of locations.</p><p><strong>That sounds like a lot indeed. What were the good parts of the traveling, I bet it wasn&apos;t only extra work?</strong></p><p>The opportunity to live in 14 cities, and stay in ~46 more, it was amazing. And traveling by land was superb compared to flying: you get to see and experience so much more! Even though I got to spend a short time in each place, the experience differs greatly from a touristy visit. For example, I liked Paris much better now when I shortly lived there compared to a long weekend visit as a tourist. But people are the most important, like always. Regardless of my time working or cycling, I had plenty of opportunities to visit friends and business associates across Europe. I also encountered like-minded people spontaneously, both remote workers and hosts related to accommodations, and cyclists on the road.  </p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2023/04/image-3.png" class="kg-image" alt="Remote-remote working" loading="lazy" width="1810" height="1113" srcset="https://blog.emblica.fi/content/images/size/w600/2023/04/image-3.png 600w, https://blog.emblica.fi/content/images/size/w1000/2023/04/image-3.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2023/04/image-3.png 1600w, https://blog.emblica.fi/content/images/2023/04/image-3.png 1810w" sizes="(min-width: 720px) 720px"><figcaption><span style="white-space: pre-wrap;">If your work enables you to cycle a 20km downhill to a sunset on the Mediterranean shore, using that opportunity can be worth it. Murcia, Spain.</span></figcaption></figure><p><strong>What advice would you give someone planning a longer stay outside their standard remote working location?</strong></p><p>Oh, there are many things. First of all, communication: discuss your plan with your supervisor and team long before you&apos;re about to leave. It is much easier to get people to agree to non-conventional plans when you show that you work well remotely and the communication between you and your team is fluent, regardless of the location.<br><br>Cybersecurity is another topic that is often overlooked. Have your devices encrypted, password manager accessible from several devices, and two-factor authentication in use. We provide opsec training for our employees and ensure that our customer&apos;s information is secure where ever we work. In computer security, the human is the weakest link. Make sure that you&apos;re not the weak link that breaks. Then people let you also travel with work devices.<br><br>A huge topic is also the climate impact of your travel. The bare minimum is always to be aware of how much emissions your travel roughly produces and consciously decide that you will cause those emissions. There are plenty of emission estimation apps, personally I trust <a href="https://www.compensate.com/individuals">Compensate app</a> to be on the safe side of the estimations.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2023/04/image-5.png" class="kg-image" alt="Remote-remote working" loading="lazy" width="1007" height="944" srcset="https://blog.emblica.fi/content/images/size/w600/2023/04/image-5.png 600w, https://blog.emblica.fi/content/images/size/w1000/2023/04/image-5.png 1000w, https://blog.emblica.fi/content/images/2023/04/image-5.png 1007w" sizes="(min-width: 720px) 720px"><figcaption><span style="white-space: pre-wrap;">Everything an IT consultant needs for half a year fits in a surprisingly small space. The work equipment, emergency shelter &amp; first aid (for cycle and the cyclist), water, food, and bike locks in the Las Bardenas desert in Northern Spain.</span></figcaption></figure><p>Lastly: remote-remote working is easier than you think. You don&apos;t <em>need</em> so many things as you might first assume, and with mobile phones, international internet coverage, and all the apps traveling is almost too easy. Yes, you need to be able to lead your own work and carry some responsibility over where your next food will manifest itself, but the opportunity to experience other places is very much worth it. If you work in IT, most probably the biggest obstacle in front of your remote-remote work experience is you and your own routines. Although most employers have restrictive policies on where and how much you are allowed to work abroad, traveling by land a few hours to a different city for a week or two is within the reach of most IT professionals.</p><p><strong>Would you say that everyone should work remotely?</strong></p><p>Absolutely not, there are a lot of aspects that you can&apos;t replace when working remotely. Especially when you don&apos;t know your colleagues well, getting to know them in person is crucial for communication and work dynamics. That&apos;s why we have hybrid roles in Emblica. And this 5-months-of-remote was a rather extreme experiment in many ways. That said, I would recommend everyone to try working remotely if possible and organize a remote-remote working experience once the basics are in place. </p><p><strong>Those are wise words; thanks for sharing your experience!</strong><br><br>Thank you for taking the time!</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2023/04/image-4.png" class="kg-image" alt="Remote-remote working" loading="lazy" width="633" height="928" srcset="https://blog.emblica.fi/content/images/size/w600/2023/04/image-4.png 600w, https://blog.emblica.fi/content/images/2023/04/image-4.png 633w"><figcaption><span style="white-space: pre-wrap;">The 5000km mark was reached just before the rock of Gibraltar, the United Kingdom.</span></figcaption></figure><p>Anna: It&apos;s great that everyone in Emblica can work in a hybrid model here. It is essential to meet your colleagues face to face, but also as an employer, to make remote working easy when possible, as some of us need quiet days at home or want to stay longer in the summer cottage or abroad. If there is an option to choose where you can work from, and it&apos;s planned well, I believe it dramatically impacts work satisfaction! </p><p>In many ways, this was an exciting and even successful experiment. Although, I am not sure if anyone will repeat the exact thing soon. Nevertheless, at Emblica, we want to support our employees to have ways of working that suit them, naturally prioritizing our customers&apos; needs and team dynamics.</p><hr><p><em>Emblica is a </em><a href="https://emblica.com"><em>data and AI consultancy</em></a><em> solving real-world problems with a large toolbox of data-relevant technologies. We work on-site and remotely, prioritizing our customers&apos; needs while pushing the limits of ways of working to suit our employees&apos; preferences.</em><br></p>]]></content:encoded></item><item><title><![CDATA[Datatieteilijän sankarimyytti]]></title><description><![CDATA[Datatieteilijät ovat nykypäivän supersankareita, jotka visioivat suuria ja pelastavat mikä tahansa liiketoiminnan haasteen pyöräyttämällä dataa notebookeissa.  Kuulostaako uskottavalta? Lue tästä mitkä ovat datatieteilijöiden jalustalle nostamisen ongelmat, ja miten ne vältetään. ]]></description><link>https://blog.emblica.fi/datatieteilijan-sankarimyytti/</link><guid isPermaLink="false">636a3134e527370001e8c1e1</guid><category><![CDATA[datatiede]]></category><dc:creator><![CDATA[Teemu Heikkilä]]></dc:creator><pubDate>Fri, 24 Mar 2023 08:30:10 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2021/03/emblicaAjattelee1-1.jpg" medium="image"/><content:encoded><![CDATA[<img src="https://blog.emblica.fi/content/images/2021/03/emblicaAjattelee1-1.jpg" alt="Datatieteilij&#xE4;n sankarimyytti"><p><em>Shortly in English: Data scientists are today&apos;s superheroes who envision big and save any business challenge by spinning some data in notebooks. Sound plausible? Read here what are the problems of putting data scientists on a pedestal, and how to avoid them.</em></p><p>Analytiikka, big data, teko&#xE4;ly, AI, data science... paljon sanoja, viel&#xE4; enemm&#xE4;n hype&#xE4;, sinisi&#xE4; s&#xE4;hk&#xF6;aivoja, robotteja ja avaruusaluksia. Olemme kehitt&#xE4;m&#xE4;ss&#xE4; liiketoimintaa alueelle jossa fantasioidaan utuisista utopioista. Samaan aikaan todella iso osa yrityksist&#xE4; haluaa esiinty&#xE4; edell&#xE4;k&#xE4;vij&#xF6;in&#xE4; ja tarjoilee yksinkertaisimmatkin algoritminsa samalla kiilt&#xE4;v&#xE4;ll&#xE4; kuvastolla. Olemmeko siis jo saavuttaneet haavekuvat vai onko todellisuus niist&#xE4; viel&#xE4; kaukana?</p><p>Moni kirjoitus muistuttaa, ett&#xE4; teko&#xE4;ly&#xE4; ja teko&#xE4;lyksi miellett&#xE4;vi&#xE4; j&#xE4;rjestelmi&#xE4; on jo yhteiskunnassamme k&#xE4;yt&#xF6;ss&#xE4; laajalti. N&#xE4;iden ratkaisujen todellisesta toiminnasta, rakenteesta ja hy&#xF6;dyist&#xE4; puhutaan kuitenkin vain v&#xE4;h&#xE4;n, mink&#xE4; vuoksi ei ole ihme ett&#xE4; &#xE4;&#xE4;nek&#xE4;s teko&#xE4;lyhype t&#xE4;ytt&#xE4;&#xE4; ymm&#xE4;rryksen tyhji&#xF6;t&#xE4; utopistisilla lupauksilla. T&#xE4;m&#xE4;n seurauksena datatiede n&#xE4;ytt&#xE4;ytyy tavalliselle kansalaiselle l&#xE4;hes mustana magiana, joka on ns. &quot;tavan tallaajan&quot; ulottumattomissa.</p><h3 id="datatiede-on-nostettu-niin-korkealle-jalustalle-ett-se-alkaa-muistuttaa-norsunluutornia">Datatiede on nostettu niin korkealle jalustalle, ett&#xE4; se alkaa muistuttaa norsunluutornia</h3><p>Miten t&#xE4;m&#xE4; hehkutus ja teko&#xE4;lyn mystifiointi on vaikuttanut ammattiaan harjoittaviin datatieteilij&#xF6;ihin? Ensinn&#xE4;kin, moderni datatieteilij&#xE4; on saavuttanut l&#xE4;hes myyttisen imagon. Ty&#xF6;paikkoja on enemm&#xE4;n kuin tekij&#xF6;it&#xE4;, kompensaatio ty&#xF6;st&#xE4; on kova ja ty&#xF6;teht&#xE4;v&#xE4;t mielenkiintoisia. Moni juuri aloittanut datatieteilij&#xE4; ajattelee tulevan uransa koostuvan sankarimaisesta ongelmien ratkomisesta ja t&#xE4;ydellisten ratkaisujen tuottamisesta yhdess&#xE4; samalla tavoin ajattelevan eliittikollegion kanssa. Jokainen toki huomaa, ett&#xE4; todellisuus on ehk&#xE4; hieman ihannettaan karumpi, mutta sankari-datatieteilij&#xE4;n ihanne n&#xE4;ytt&#xE4;&#xE4; muovanneen alaa salakavalasti.</p><figure class="kg-card kg-image-card kg-width-wide kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2021/03/Emblicaview2.jpg" class="kg-image" alt="Datatieteilij&#xE4;n sankarimyytti" loading="lazy" width="2000" height="1334" srcset="https://blog.emblica.fi/content/images/size/w600/2021/03/Emblicaview2.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2021/03/Emblicaview2.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2021/03/Emblicaview2.jpg 1600w, https://blog.emblica.fi/content/images/2021/03/Emblicaview2.jpg 2000w" sizes="(min-width: 1200px) 1200px"><figcaption>Sankari-eliittiin kuuluva datatieteilij&#xE4; visioi tulevaisuuden ratkaisuja norsunluutornissaan.</figcaption></figure><p>Harmillisen suuri osa suomalaisten organisaatioiden datatieteilij&#xF6;ist&#xE4; onkin j&#xE4;rjest&#xE4;ytynyt t&#xE4;t&#xE4; sankari-eliitti&#xE4; mukaillen. Siiloutunut, pieni porukka pohtii ongelmia usein erillisess&#xE4; data- tai analytiikkatiimiss&#xE4;, ik&#xE4;&#xE4;n kuin omana supersankarien joukkonaan, joka rient&#xE4;&#xE4; tarvittaessa organisaatiota painavien dataongelmien kimppuun. T&#xE4;m&#xE4; asetelma luo pohjan monenlaisille ongelmille:</p><ol><li>Ensimm&#xE4;inen ongelma on kulttuurinen: <strong>datatieteilij&#xF6;iden eliittiporukka yksinkertaisesti &#xE4;rsytt&#xE4;&#xE4;, eik&#xE4; sooloilu tuo projekteihin toimivia lopputuloksia.</strong> Digitalisaatio on pakottanut ohjelmistokehitt&#xE4;j&#xE4;t eritt&#xE4;in tiiviiseen yhteisty&#xF6;h&#xF6;n liiketoiminnan kanssa, ja onkin ilahduttavaa n&#xE4;hd&#xE4; kuinka sujuvasti Suomessa tehd&#xE4;&#xE4;n nyky&#xE4;&#xE4;n monialaisia softaprojekteja. Jostain syyst&#xE4; datatiimit n&#xE4;hd&#xE4;&#xE4;n kuitenkin t&#xE4;st&#xE4; erillisen&#xE4; porukkana. Kun kehitystiimiin tuodaan dataosaamista, painajaisten &#x201C;eliitti-tiimi&#x201D; vaatii p&#xE4;&#xE4;sy&#xE4; kaikkeen dataan ja tuuppaa kehitystiimin backlogille valtavasti teht&#xE4;vi&#xE4; liittyen esimerkiksi datan k&#xE4;sittelyyn. Lopputuloksena kaikesta vaivasta on kasa hankalasti integroitavia Python-notebookkeja ja lupaus paremmasta tulevaisuudesta. Harmillisesti syvemm&#xE4;lle yhteisty&#xF6;lle ei j&#xE4;&#xE4;nyt kuitenkaan aikaa, koska dataporukka on jo kutsuttu seuraaviin seikkailuihin.</li><li><strong>Toinen ongelma liittyy datatiimien diversiteetin puutteeseen.</strong> Koodareiden diversiteettiongelmista on keskusteltu i&#xE4;t ja ajat, mutta datatieteilij&#xE4;t voitaisiin lis&#xE4;t&#xE4; tilanteen polttavuuden vuoksi ongelmalistan k&#xE4;rkeen. Vaikka hyv&#xE4; tiimi osaa toki ottaa ulkopuolelta tulevia vaatimuksia ja n&#xE4;kemyksi&#xE4; huomioon, on tiivis samoin ajatteleva porukka kuitenkin pohjimmiltaan kaikukammio - ja se n&#xE4;kyy. Ratkaisut rakennetaan yhdenlaiseen maailmankuvaan ja moraalik&#xE4;sitykseen perustuen, eik&#xE4; toisenlaisia n&#xE4;k&#xF6;kulmia tajuta ajatella tarpeeksi laajasti. Edell&#xE4;mainittu erist&#xE4;ytyminen erikoisryhm&#xE4;ksi ei tietenk&#xE4;&#xE4;n helpota t&#xE4;t&#xE4; tilannetta, vaan pahentaa kuilua &#x201C;eliitti-tiimin&#x201D; sek&#xE4; moninaisten sidosryhmien v&#xE4;lill&#xE4;.</li><li><strong>Datatieteilij&#xF6;iden diversiteettiongelmaa voi laajentaa my&#xF6;s ty&#xF6;menetelmiin.</strong> Jostain syyst&#xE4; nimitt&#xE4;in tuntuu, ett&#xE4; ohjelmistokehityksen menetelm&#xE4;t koetaan jotenkin datatieteest&#xE4; irrallisena asiana, vaikka todellisuudessa datatieteilij&#xF6;iden menetelm&#xE4;t eiv&#xE4;t yksin&#xE4;&#xE4;n tuota suurtakaan arvoa, jos dataprojektien lopputulokset j&#xE4;&#xE4;v&#xE4;t ohjelmistoratkaisusta irrallisiksi. Eripuraakin syntyy kun koetaan, ett&#xE4; kalliiden &#x201C;eliitti-aivojen&#x201D; ty&#xF6;aikaa ei kannata tuhlata dataputken hiomiseen, eik&#xE4; siihen oikeastaan ole kiinnostustakaan. Versionhallinta, deploy, devops ja testaus ovat normaali osa jokaista ohjelmistoprojektia, eik&#xE4; datatieteilij&#xF6;iden &#x201C;magialla&#x201D; p&#xF6;tkit&#xE4; datan hy&#xF6;dynt&#xE4;misess&#xE4; pitk&#xE4;lle, mik&#xE4;li n&#xE4;m&#xE4; vaiheet puuttuvat.</li></ol><h3 id="kuinka-v-ist-sankarimyytin-tuomat-ongelmat">Kuinka v&#xE4;ist&#xE4;&#xE4; sankarimyytin tuomat ongelmat?</h3><figure class="kg-card kg-image-card kg-width-wide kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2021/03/EmblicaAjattelee3.jpg" class="kg-image" alt="Datatieteilij&#xE4;n sankarimyytti" loading="lazy" width="2000" height="1334" srcset="https://blog.emblica.fi/content/images/size/w600/2021/03/EmblicaAjattelee3.jpg 600w, https://blog.emblica.fi/content/images/size/w1000/2021/03/EmblicaAjattelee3.jpg 1000w, https://blog.emblica.fi/content/images/size/w1600/2021/03/EmblicaAjattelee3.jpg 1600w, https://blog.emblica.fi/content/images/2021/03/EmblicaAjattelee3.jpg 2000w" sizes="(min-width: 1200px) 1200px"><figcaption>Datatieteilij&#xE4;t ty&#xF6;skentelev&#xE4;t samassa ryhm&#xE4;ss&#xE4; muun devaustiimin kanssa.</figcaption></figure><ol><li><strong>Datatieteilij&#xE4;t eiv&#xE4;t vain suunnittele, vaan toteuttavat softaa yhdess&#xE4; muun tiimin kanssa.</strong> Mink&#xE4; tahansa dataprojektin kannalta on hy&#xF6;dyllist&#xE4;, ett&#xE4; datatieteilij&#xE4;t ja -insin&#xF6;&#xF6;rit kykenev&#xE4;t suoriutumaan datan hy&#xF6;dynt&#xE4;misen lis&#xE4;ksi my&#xF6;s projektissa vaadittavasta perinteisest&#xE4; ohjelmistokehityksest&#xE4;. T&#xE4;llaiseen ty&#xF6;skentelyyn kuuluu ymm&#xE4;rrys siit&#xE4;, ett&#xE4; koko tiimi rakentaa yhdess&#xE4; softaa johon tuodaan lis&#xE4;arvoa datalla. Onnistumisen kannalta kaikki toteutettava ty&#xF6; onkin siis arvokasta, ja koko tiimi kantaa siit&#xE4; yhteisen vastuun.</li><li><strong>Projektien omistajuus kuuluu alusta asti koko tiimille, etenkin asiakkaalle.</strong> </li><li><strong>Tiimin diversiteetin puuttuminen pit&#xE4;&#xE4; tiedostaa ja asian korjaamiseksi on teht&#xE4;v&#xE4; t&#xF6;it&#xE4;.</strong> On hyvin loogista, ett&#xE4; erilaisista taustoista tulevat ihmiset onnistuvat yhdess&#xE4; ymm&#xE4;rt&#xE4;m&#xE4;&#xE4;n laajempia k&#xE4;ytt&#xE4;j&#xE4;ryhmi&#xE4; sek&#xE4; k&#xE4;ytt&#xF6;tapauksia, jonka vuoksi heill&#xE4; on suurempi todenn&#xE4;k&#xF6;isyys onnistua &#xE4;lykk&#xE4;&#xE4;ss&#xE4;, useita k&#xE4;ytt&#xF6;tapauksia palvelevassa datan hy&#xF6;dynt&#xE4;misess&#xE4;. Siksi toimiva datatiimi saavutetaan palkkaamalla ihmisi&#xE4; erilaisista taustoista.</li><li><strong>Datan hy&#xF6;dynt&#xE4;misen tueksi tarvitaan ymm&#xE4;rryst&#xE4; bisneksen erityisominaisuuksista.</strong> Dataprojekteja ajaa l&#xE4;hes aina taloudellisen kasvun tai s&#xE4;&#xE4;st&#xF6;jen tavoittelu, eik&#xE4; projekteja voi siksi suunnitella tai toteuttaa ymm&#xE4;rt&#xE4;m&#xE4;tt&#xE4; kunkin liiketoiminta-alueen lainalaisuuksia. Meid&#xE4;n kannaltamme on ollut hy&#xF6;dyllist&#xE4;, ett&#xE4; l&#xE4;hes kaikissa projekteissamme on mukana palvelumuotoilua, joka mahdollistaa liiketoiminnasta keskustelun kokonaisuutena. Pelkk&#xE4;&#xE4;n teknologiaan keskittyminen tarjoaa mahdollisuuden syv&#xE4;&#xE4;n erikoistumiseen, mutta liiketoimintaymm&#xE4;rryksen tuominen osaksi projektien toteutusta tekee edistyneist&#xE4; ideoista oikeasti hy&#xF6;dyllisi&#xE4; ja k&#xE4;ytt&#xF6;kelpoisia.</li><li><strong>Emme toimita pelkki&#xE4; Python-notebookeja (paitsi jos niit&#xE4; hyv&#xE4;st&#xE4; syyst&#xE4; pyydet&#xE4;&#xE4;n).</strong> Ty&#xF6;tapojen monipuolisuuden osalta meill&#xE4; on pienest&#xE4; tiimist&#xE4; huolimatta laajasti kokemusta ja ambitioita erilaisia teknologioita ja ongelmia kohtaan, emmek&#xE4; j&#xE4;t&#xE4; t&#xF6;it&#xE4;mme nopean &#x201C;datapy&#xF6;r&#xE4;ytyksen&#x201D; tasolle. Vaativatpa projektit sitten arkkitehtuurin suunnittelua, syvi&#xE4; neuroverkkoja, optimointi- tai kombinatoristen ongelmien ratkaisua, tiimist&#xE4;mme l&#xF6;ytyy siihen sek&#xE4; kiinnostusta ett&#xE4; osaamista.</li></ol><p>Datatieteen tilanne ei ole miss&#xE4;&#xE4;n nimess&#xE4; menetetty, mutta juuri siksi aiheesta on syyt&#xE4; puhua. Ratkaisu ongelmaan l&#xF6;ytyy toimintatapojen ja tiimien diversiteetist&#xE4;, kollegoiden v&#xE4;lisest&#xE4; yhteispelist&#xE4; ja projektien laajasta omistajuudesta. N&#xE4;in rakennetaan onnistuneita dataprojekteja, joilla on kyky kasvattaa liiketoimintaa ja tuoda muutakin lis&#xE4;arvoa yhteiskuntaamme. Vastuu ja mahdollisuus vaikuttaa on meill&#xE4;, datan ammattilaisilla.</p><p>Etsimme jatkuvasti uusia tekij&#xF6;it&#xE4;, jotka eiv&#xE4;t pelk&#xE4;&#xE4; tuoda n&#xE4;kemyksi&#xE4;&#xE4;n julki. Tutustu meihin ty&#xF6;nantajana t&#xE4;&#xE4;ll&#xE4;: <a href="https://emblica.com/careers">emblica.com/careers</a> </p><hr><p><em><em><a href="https://emblica.com/">Emblica</a> ei ole se tavallinen datatiimi. Rakennamme r&#xE4;&#xE4;t&#xE4;l&#xF6;ityj&#xE4; ratkaisuja datan ker&#xE4;&#xE4;miseen, k&#xE4;sittelyyn ja hy&#xF6;dynt&#xE4;miseen alalle kuin alalle, etenkin R&amp;D:n rajapinnassa. Oli kohteemme tehdaslinjasto, verkkokauppa tai pelto, l&#xF6;yd&#xE4;t meid&#xE4;t ty&#xF6;n touhusta, k&#xE4;det savessa - ainakin toimistoltamme Helsingist&#xE4;.</em></em></p>]]></content:encoded></item><item><title><![CDATA[There Is More To Computer Vision Than Video Cameras]]></title><description><![CDATA[What advantages does millimeter-wave radar offer in perception problems where visibility and privacy are limited? Read how we at Emblica have utilized mmWave technology to solve problems where camera-based computer vision solutions are not viable.]]></description><link>https://blog.emblica.fi/there-is-more-to-computer-vision-than-video-cameras/</link><guid isPermaLink="false">63ecbcc049d0670001f8520b</guid><category><![CDATA[Computer Vision]]></category><category><![CDATA[radar]]></category><category><![CDATA[mmWave radar]]></category><dc:creator><![CDATA[Samuel Piirainen]]></dc:creator><pubDate>Mon, 20 Feb 2023 19:25:04 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2023/02/mmWave-radar-image--1-.png" medium="image"/><content:encoded><![CDATA[<img src="https://blog.emblica.fi/content/images/2023/02/mmWave-radar-image--1-.png" alt="There Is More To Computer Vision Than Video Cameras"><p><em>Shortly in Finnish: Mit&#xE4; etuja millimetriaaltotutka tarjoaa havainnointiongelmissa, joissa n&#xE4;kyvyydess&#xE4; on puutteita tai yksityisyydensuoja on ensisijaista? Lue t&#xE4;st&#xE4;, miten me Emblicalla olemme ratkaisseet mmWave-tekniikan avulla ongelmia, joissa kameraa hy&#xF6;dynt&#xE4;v&#xE4;t konen&#xE4;k&#xF6;ratkaisut eiv&#xE4;t ole olleet mahdollisia.</em></p><!--kg-card-begin: markdown--><p>For the past several years, automation and AI have had great levels of interest in almost any industry from car R&amp;D to agriculture <sup><a href="https://ieeexplore.ieee.org/abstract/document/9016391">1</a> </sup> <sup> <a href="https://www.sciencedirect.com/science/article/pii/S2589721719300182">2</a></sup>. <i>Computer Vision</i> (CV) in particular is a domain that continues to find countless applications via Machine Learning. However, these solutions largely rely on Camera-based computer vision&#x2014;perhaps as it is the most intuitive kind of vision to us humans. At Emblica, we&apos;re comfortable using also less common tools and adopting a technology that is not common knowledge even in the domain of Machine Learning specialists we operate in. This can provide our customers unique solutions to problems otherwise deemed too difficult to solve. In this blog, we focus on one alternative to Camera-based Computer Vision, the millimeter-wave radar.</p>
<h2 id="limitations-of-camera-based-computer-vision">Limitations of Camera-based Computer Vision</h2>
<p>As mentioned, camera-based systems are not always suitable for CV applications. Here are some reasons why:</p>
<ol>
<li>
<p><b>Privacy concerns</b>: Despite the availability of technology to blur faces, privacy remains a concern. There is always the risk of malicious third-party access to raw video data, even if it is anonymized in post-processing.</p>
</li>
<li>
<p><b>Environmental issues</b>: Cameras are susceptible to weather conditions, such as fog or dirt on the lens. This makes designing camera-based systems more complicated.</p>
</li>
<li>
<p><b>Hardware requirements</b>: The specific hardware requirements for camera-based systems can be challenging.</p>
</li>
</ol>
<h2 id="alternatives-to-cameras">Alternatives to Cameras</h2>
<p>In light of the challenges posed by camera-based computer vision solutions, it is important to consider alternative technologies that can address these issues. One such technology is radars, which offer several advantages over cameras. For instance, radar signals cannot be used to identify facial features, thus addressing privacy concerns. Additionally, radars can penetrate through sparse matter, allowing for clear radar images even in low-visibility conditions like dust, rain, or snowfall. Furthermore, radars do not have a camera lens, eliminating the risk of obstructions in the field of view. These features make radars a promising technology for many special Computer Vision applications where cameras are inadequate.</p>
<p>One well-known technology&#x2014;in many ways analogous to radars&#x2014;are LiDAR-based systems. While Devices utilizing the LiDAR-technology have found a plenty of use in many industrial applications like self-driving cars <sup><a href="https://opg.optica.org/directpdfaccess/ac471164-8be8-4760-9e28174371ac3f7e_380434/opn-29-1-26.pdf">3</a></sup>. For those interested to read more about LiDAR-technology, here is <a href="https://blog.emblica.com/points-about-point-clouds/">our blog post on the topic</a>. However, a downside of the technology is these devices are very expensive. This makes them impractical for many applications where the system must be affordable for mass production.</p>
<p><i>A millimeter-wave radar</i> (mmWave radar) is a small, relatively cheap (available at prices well under 200 euro at the time of publication) radar that operates on a very precise millimeter-wave signal. It is precise enough that it is possible to discover the position and gait of a person even through sparse matter&#x2014;such as obstructions on the lens&#x2014;without discovering the facial features.</p>
<h2 id="what-is-a-millimeter-wave-radar">What is a Millimeter-Wave Radar?</h2>
<p>A mmWave radar is a type of radar that emits short-wavelength electronic waves known as millimeter-waves (mmWaves), that objects in their path then reflect back <sup> <a href="http://www.ee.fju.edu.tw/images/file/data3169/9a41b2657fe717233110268d676d38ef.pdf">4</a></sup>. The radar captures these reflected signals, and the underlying software of the system can usually calculate properties such as the range, relative radial velocity, and angle of arrival of targets within the signal. These are physical properties traditional to a radar-system, usually obtained via signal processing. This is in contrast to a camera, whose signal only reveals the 2D position.</p>
<p>An interesting property of the mmWave-signal in particular is that its wavelength in the electromagnetic spectrum is short. This has two advantages: Firstly, the radar system can be relatively small in size.  Secondly, the accuracy of the signal is high. For example, according to Texas Instruments, a mmWave radar operating at 76&#x2013;81 GHz can detect movements as small as a fraction of a millimeter <sup><a href="http://www.ee.fju.edu.tw/images/file/data3169/9a41b2657fe717233110268d676d38ef.pdf">4</a></sup>.</p>
<figure style="margin: 40px 0px 40px 0px;">
  <img style="margin-bottom:15px;" alt="There Is More To Computer Vision Than Video Cameras" src="https://blog.emblica.fi/content/images/2023/02/Screenshot-2023-02-15-at-13.49.24.png" width="600">
  <figcaption style="font-size:14px;">A top-down view of two humans observed in the mmWave radar signal.</figcaption>
</figure>
<blockquote>
<p>Because the domain is very different to camera-based computer vision, discovering important information from signals like mmWave-signal requires sophisticated technology and specialized expertise. Machine Learning tools can be a great help here.</p>
</blockquote>
<p>Moreover, many off-the-shelf systems of  mmWave radars come with firmware for transforming the radar signal into a three-dimensional point cloud. Once objects are observed in the vicinity, their position and relative radial velocity can be seen in the point cloud data output by the system. This is also where Machine Learning and AI come into picture. Because the domain is very different to camera-based computer vision, discovering important information from signals like mmWave-signal requires sophisticated technology and specialized expertise. Machine Learning tools can be a great help here.</p>
<h2 id="millimeter-wave-radar-applications">Millimeter-Wave Radar Applications</h2>
<p>Millimeter-wave radar has numerous potential applications. One such use case is the detection and counting of humans within the radar&apos;s range. For instance, smart doors or elevators that can distinguish between someone simply walking by and someone waiting for the door to open. Another potential application is detecting falls among elderly residents in nursing homes, which can be crucial for their well-being and health. While difficult problems for classical algorithms, such tasks can be automated with modern Machine Learning tools, thanks to their capability to deal with large masses of data, without compromising the individuals&apos; privacy.</p>
<blockquote>
<p>While difficult problems for classical algorithms, such tasks can be automated with modern Machine Learning tools, thanks to their capability to deal with large masses of data, without compromising the individuals&apos; privacy.</p>
</blockquote>
<blockquote>
<p>Even outside cars, autonomous vehicles and machines are prime example of environment where radar can be used and where some of our customers are developing new solutions.</p>
</blockquote>
<p>Object detection solutions with the mmWave radar are more common than one might think. The radar is used by several car manufacturers (e.g. Tesla, Mercedes-Benz, and Audi <sup><a href="https://www.researchgate.net/publication/353766805_MmWave_Radar_and_Vision_Fusion_based_Object_Detection_for_Autonomous_Driving_A_Review">5</a></sup>) in sensor fusion combined with other radar types like LiDARs to improve the capabilities of the cars&apos; autonomous driving systems to make decisions.</p>
<p>Even outside cars, autonomous vehicles and machines are prime example of environment where radar can be used and where some of our customers are developing new solutions.</p>
<blockquote>
<p>mmWave radars have been used to measure heart beats and breathing remotely.</p>
</blockquote>
<p>Advantages in comparison to using just the more well-known LiDARs include improved long-range object detection, lower price, and more accurate detection of dynamic targets. In comparison to vision sensors, the object detection performance of mmWave radars is also  affected less by extreme weather.</p>
<p>The possibilities do not end with object detection, mmWave radars have been used to measure heart beats and breathing remotely and with sophisticated algorithms object identification from the signal is also viable. Maybe we would like to know how many shopping mall customers prefer to use shopping carts or hand-held baskets? The difference in the objects can be seen in the mmWave-signal, and extracted using Machine Learning tools.</p>
<h2 id="human-identification-and-tracking-in-indoor-environments">Human Identification and Tracking in Indoor Environments</h2>
<p>At Emblica, we have investigated the viability of the mmWave radar in the problem domain of human detection and tracking in an indoor space. In particular, we investigated the Emblica office and the public transport. These are good examples of environments where the mmWave radar has an advantage: (1) The spaces are in the public, so anonymity has a high priority. (2) Public surfaces are prone to getting dirty and degraded, either due to natural wear and tear or due to harassment.</p>
<figure style="margin: 40px 0px 40px 0px;">
  <img style="margin-bottom:15px;" alt="There Is More To Computer Vision Than Video Cameras" src="https://blog.emblica.fi/content/images/2023/02/setup--1-.jpeg" width="300">
  <figcaption style="font-size:14px;">MMWR-Tracker system data collection setup in Emblica premises.</figcaption>
</figure>
<p>This task can be reduced to a problem of tracking humans in an indoor space. To this end, Emblica engineered the <i>MMWR-Tracker</i>&#x2014;an end-to-end system for human detection and tracking in an indoor environment.</p>
<figure style="margin: 40px 0px 40px 0px;">
  <img style="margin-bottom:15px;" alt="There Is More To Computer Vision Than Video Cameras" src="https://blog.emblica.fi/content/images/2023/02/flow_blog.png" width="600">
  <figcaption style="font-size:14px;">The MMWR-Tracker system architecture.</figcaption>
</figure>
<p>The system uses the U-Net architecture <sup><a href="https://arxiv.org/pdf/1505.04597.pdf">6</a></sup> &#x2014;A Deep Neural Network for Image Segmentation applications&#x2014;to first segment the mmWave-signal from the mmWave radar system into a segmentation map where signal noise is filtered away and humans detected by the signal remain. These segmentation maps are passed to a tracking module, that utilizes a Kalman filter and the Hungarian Method&#x2014;methods commonly found in object tracking solutions in the domain&#x2014;to track the humans detected.</p>
<figure style="margin: 40px 0px 40px 0px; display: flex; align-items: center;">
    <div style="display: inline-block; width: 48%; margin: 1%;">
          <img style="margin-bottom:20px;" alt="There Is More To Computer Vision Than Video Cameras" src="https://blog.emblica.fi/content/images/2023/02/setup_bus3--1--1.jpeg" width="300">
        <figcaption style="font-size:14px;">Data collection within a bus.</figcaption>
    </div>
        <div style="display: inline-block; width: 48%; margin: 1%;">
          <img style="margin-bottom:20px;" alt="There Is More To Computer Vision Than Video Cameras" src="https://blog.emblica.fi/content/images/2023/02/setup_bus2.jpeg" width="300">
            <figcaption style="font-size:14px;">The mmWave radar directed towards the bus aisle.</figcaption>
    </div>
</figure>
<p>Once installed in an <i>edge device backend</i>, the MMWR-Tracker works in real-time, monitoring the environment and relaying the spatial information of detected objects to the cloud.</p>
<p>Tracking humans is just one possible application of the radar technology. Other possibilities include detecting whether a human or a pet is moving in its field of view&#x2014;think of smart lighting or security systems.</p>
<h2 id="conclusion">Conclusion</h2>
<p>In conclusion, while many computer vision solutions rely on cameras, the mmWave radar offers an alternative solution that is especially useful in scenarios where visibility and privacy are limited. Although not as well-known as other computer vision tools, such as stereo cameras or LiDAR systems, the mmWave radar has unique properties that make it a valuable consideration for specific use cases. Emblica has explored the potential of this technology to address problems that cannot be solved with camera-based computer vision solutions.</p>
<!--kg-card-end: markdown--><hr><p><em><a>Emblica</a></em> <em>is not your average data team. We build customized solutions for collecting, processing, and utilizing data for all sectors, especially at the R&amp;D interface. Whether our target is a factory line, an online store, or a bus, you can find us busy at work, hands in the clay - at least at our office in Helsinki.</em></p>]]></content:encoded></item><item><title><![CDATA[Points about point clouds]]></title><description><![CDATA[<p>Representing 3D space digitally can be a challenging task. The most common way to process visuals about our world comes in the form of a 2D image, either from a photograph or video. Capturing 3D images might be more common than you think. Sensors using technologies such as LiDAR or</p>]]></description><link>https://blog.emblica.fi/points-about-point-clouds/</link><guid isPermaLink="false">636a3561e527370001e8c24d</guid><category><![CDATA[point clouds]]></category><category><![CDATA[lidar]]></category><category><![CDATA[radar]]></category><category><![CDATA[machine learning]]></category><category><![CDATA[Machine Vision]]></category><category><![CDATA[ENG]]></category><category><![CDATA[ML]]></category><dc:creator><![CDATA[Rick Joosten]]></dc:creator><pubDate>Tue, 14 Feb 2023 08:00:00 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2022/11/colorful_pointcloud_of_the_outline_of_a_pyramid-_seen_from_0.png" medium="image"/><content:encoded><![CDATA[<img src="https://blog.emblica.fi/content/images/2022/11/colorful_pointcloud_of_the_outline_of_a_pyramid-_seen_from_0.png" alt="Points about point clouds"><p>Representing 3D space digitally can be a challenging task. The most common way to process visuals about our world comes in the form of a 2D image, either from a photograph or video. Capturing 3D images might be more common than you think. Sensors using technologies such as LiDAR or radar do this. These sensors often produce data in the form of point clouds. In this blog, we will walk you through;</p><ul><li>What point clouds are.</li><li>How point cloud data is created.</li><li>What kind of problems can be tackled using point clouds.</li></ul><p><em><strong>Lyhyesti suomeksi:</strong> Tutkat, lidarit ja muut anturit ovat kustannustehokas tapa tuottaa pistepilvi&#xE4;, eli moniulotteisia pistejukkoja. Lue t&#xE4;st&#xE4; tarkemmin, miten pistepilvidataa luodaan ja millaisia ongelmia voidaan ratkaista pistepilvien avulla.</em></p><h1 id="what-are-point-clouds">What are point clouds?</h1><p>Point clouds are simply collections of points in space. Most of the time we are talking about 3D space, but a 2D scatterplot could also be considered a point cloud. The same goes for a higher dimensional space.</p><p>There are two important characteristics of point clouds. First, a point cloud is unordered. This means that the set of points can be shuffled and put in any order and it would still be the same point cloud. In other words, the first point is unrelated to the second point. This is not the case in regular photographs, two pixels that are next to each other are related to each other. You cannot shuffle them around and still get the same picture. This difference is important to remember when discussing point cloud algorithms later.</p><p>The second characteristic is that the space in which the point cloud exists needs to have some way to calculate the distance between two points. The most commonly used space is 3D Euclidean space where each point has an x,y, and z coordinate.</p><h2 id="how-are-point-clouds-created">How are point clouds created?</h2><figure class="kg-card kg-image-card kg-width-wide"><img src="https://blog.emblica.fi/content/images/2022/11/lidar-3.png" class="kg-image" alt="Points about point clouds" loading="lazy" width="1380" height="489" srcset="https://blog.emblica.fi/content/images/size/w600/2022/11/lidar-3.png 600w, https://blog.emblica.fi/content/images/size/w1000/2022/11/lidar-3.png 1000w, https://blog.emblica.fi/content/images/2022/11/lidar-3.png 1380w" sizes="(min-width: 1200px) 1200px"></figure><p>The most common methods for generating 3D point cloud data are using LiDAR, radar, and Photogrammetry. LiDAR (<em>Light Detection and Ranging</em>) sensors use a pulsing laser to measure an object&#x2019;s distance from the lidar based on the time it takes for the light to bounce back to the LiDAR unit. Combining this with the location of the sensor, and its directionality, the 3D location from where the pulse was reflected can be calculated. &#xA0;The LiDAR sensor can be static on the ground or attached to a moving object such as a plane or a car. In the latter case, extra care has to be taken to know the exact location and direction of the sensor. These days even modern iPhones come with a built-in Lidar</p><figure class="kg-card kg-image-card kg-width-wide kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2022/11/iphonelidar.png" class="kg-image" alt="Points about point clouds" loading="lazy" width="1920" height="1080" srcset="https://blog.emblica.fi/content/images/size/w600/2022/11/iphonelidar.png 600w, https://blog.emblica.fi/content/images/size/w1000/2022/11/iphonelidar.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2022/11/iphonelidar.png 1600w, https://blog.emblica.fi/content/images/2022/11/iphonelidar.png 1920w" sizes="(min-width: 1200px) 1200px"><figcaption>Creating a 3D image using the iPhone&apos;s built-in LiDAR sensor resulting in a 3d image of the scanned object.</figcaption></figure><p>Oftentimes a color value as well as an intensity measure of how much of the laser beam is being measured for each point. Combining these with the depth information creates richer data and can increase the usefulness for solving machine learning problems. &#xA0; <br><br>Similarly to LiDAR sensors, radar can be used to create point cloud data. Instead of light, radar uses electromagnetic waves to detect objects. This can be beneficial compared to LiDAR sensors when detecting is necessary through objects that block light but not radio waves. One place where both light and radio waves cannot penetrate far enough is underwater. <a href="https://arxiv.org/pdf/2205.09361.pdf">Here sonar could be used</a> to create point clouds using the same principles as LiDAR or radar.</p><p>The last common method is photogrammetry. This method creates a 3D image by interpreting multiple 2D images of an object. This is analogous to how humans see in 3D. These images usually have a much lower resolution compared to the previously discussed methods but are generally much cheaper to create. For example, Google Earth uses photogrammetry to create a 3D image of buildings based on satellite images.</p><h1 id="what-machine-learning-tasks-can-we-do-on-point-clouds">What machine learning tasks can we do on point clouds?</h1><p>There is a point cloud, now what? Most problems on point cloud data that we can solve with machine learning fall into one of three categories: classification, segmentation, and localization. This section will discuss each of the three problems and how to solve them.</p><h2 id="classification">Classification</h2><p>Classification is the task of detecting what object is in the point cloud. For example, products coming down a factory conveyor belt can be scanned and classified whether they are good or bad based on its shape. In this case, the point cloud only contains a single object to be classified into a class. Of course, the number of possible classes can be bigger than two.</p><h2 id="segmentation">Segmentation</h2><p>Segmentation concerns itself with dividing a point cloud into recognizable parts. This can be done at two levels. The first level is <em>part segmentation. </em>Here the task is to segment an object into recognizable parts. For instance, we know that the object is a table but find the points that are part of the top of the table and which points are part of the legs. The second level is <em>semantic segmentation.</em> Instead of segmenting a single object into its separate parts, semantic segmentation is the task of dividing a point cloud with multiple objects into distinct objects, for example, dividing a scene that is created by a self-driving car into pedestrians, other cars, buildings, busses, etc. so that the car knows what kind of objects are around. </p><h2 id="localization">Localization</h2><p>Object localization tasks are not often mentioned in the classic academic point cloud algorithms but can prove extremely useful. The goal of the task is to pinpoint a specific location on an object. This could for example be useful on a factory line where a robot has to interact with a specific part of an object. &#xA0;Each object will be the same but there might be variations in the exact location which the robot has to account for. Furthermore, if we can localize an object over time, it is possible to track the object&#x2019;s trajectory.</p><figure class="kg-card kg-image-card kg-width-wide kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2022/11/pointcloud.png" class="kg-image" alt="Points about point clouds" loading="lazy" width="2000" height="592" srcset="https://blog.emblica.fi/content/images/size/w600/2022/11/pointcloud.png 600w, https://blog.emblica.fi/content/images/size/w1000/2022/11/pointcloud.png 1000w, https://blog.emblica.fi/content/images/size/w1600/2022/11/pointcloud.png 1600w, https://blog.emblica.fi/content/images/2022/11/pointcloud.png 2244w" sizes="(min-width: 1200px) 1200px"><figcaption>Three different tasks: 1) Classification, what is the object in the point cloud? (a pyramid) 2) Segmentation, segmenting the point cloud into distinct elements 3) Localization, locating the coordinates of the pyramid&apos;s top point</figcaption></figure><h3 id="so-how-do-we-solve-these-problems-using-machine-learning">So how do we solve these problems using machine learning?</h3><p>One way is to turn to proven efficient techniques from computer vision utilizing convolutional neural networks (CNNs). From a 3D point cloud, it is possible to create a 2D depth map by essentially looking at the point cloud from a top-down birds-eye perspective. This creates an image where each pixel represents a column of the point cloud. Information about the height, number of points, reflectivity, or other characteristics can be encoded in separate channels just like a color image has a red, green, and blue channel. </p><p>The analogous structure of 2D pixels in 3D is voxels. Instead of a grid of 2D pixels, the point cloud can be represented as a grid of 3D boxes. Because the grid has a regular structure, voxels close to each other contain information about that region of the point cloud. Just like for 2D images, CNNs can be used to solve our tasks. However, the size of the image in a voxel-based approach scales cubically with the size of the point cloud compared to quadratically for 2D pixels. This makes using voxels significantly slower due to the increased overhead. </p><p>Creating a 2D picture is a good representation to use for machine learning tasks because it is a well-studied use case. Point clouds are however closer to what humans see. Machine learning using a 3D point cloud as input without processing it into a regular grid structure cannot rely on this structured nature. At the core of most solutions built upon unaltered point clouds lies PointNet (or its extension PointNet++) which deals with the unordered nature of point clouds. The next section will go deeper into the workings of this PointNet. For more information on the pixel, voxel, and other approaches check out <a href="http://vedder.io/misc/KyleVedderWPEII2021.pdf">this document</a> by Kyle Vedder.</p><h3 id="pointnet">PointNet</h3><p>The PointNet architecture is designed to create a global representation of the whole point cloud in a single vector by using only permutation invariant operations. Most importantly the aggregation function cannot rely on the order of the input. PointNet uses <em>max pooling</em> to create the global representation (no matter what the order is, the maximum will always be the same).</p><p>With the global representation, we can now add the final part of our network depending on the task we want to perform. For classification, we add a classification head that predicts to which class the whole point cloud belongs to. </p><p>If the goal is to predict the location of an object in a space we can add a regression head after the global representation. Using supervised learning methods this regression head can learn the exact location of a feature of the point cloud. For example, pinpoint the center of a wheel of a car and give its x,y, and z coordinates.</p><p>For segmentation, the model architecture can be similar to the classification architecture. However, instead of classifying the whole point cloud, segmentation requires the classification of each point. What makes segmentation tasks more challenging is that the global representation created by the PointNet might not contain enough fine-grained information about the object. One way to alleviate this problem is to concatenate the earlier layer&#x2019;s output to the global feature vector to also have information about earlier, less global, abstractions. This works reasonably well for many cases but for more challenging segmentation problems PointNet++ was created to deal with this problem.</p><p>This post will not explain PointNet++ in full detail. If you are interested, I highly recommend reading the <a href="https://arxiv.org/abs/1706.02413">original paper</a>, or <a href="https://github.com/maziarraissi/Applied-Deep-Learning/blob/main/01%20-%20Computer%20Vision/06%20-%203D.pdf">slides</a> and <a href="https://www.youtube.com/watch?v=FAqN0KK_2kg">youtube videos</a> made by Maziar Raissi about PointNet and other point cloud methods. In brief, PointNet++ clusters points in the point cloud and applies a mini-PointNet to each cluster instead of the whole point cloud. This ensures that more local information is retained and segmentation can perform better. </p><h3 id="final-thoughts">Final thoughts</h3><p>Point clouds are a great way to represent a 3D environment. Tasks that can be defined as a classification, segmentation, or localization task can be solved with the help of a PointNet. However, simply having a digital representation of a real-life environment <a href="https://www.wwf.org.uk/sites/default/files/2019-04/Lidar-WWF-guidelines.pdf">can already be beneficial</a>. For example, creating point clouds at different points in time allows for easy comparison of a 3D object. Furthermore, it is easy to zoom in on or rotate a digital object which might be difficult to do in the real world. As an additional benefit, most methods of creating point clouds are inherently more private compared to photos or videos because no personal data can easily be gathered from a point cloud.</p><hr><p><em><a href="https://blog.emblica.fi/ghost/emblica.com">Emblica</a> is not your average data team. We build customized solutions for collecting, processing, and utilizing data for all sectors, especially at the R&amp;D interface. Whether our target is a factory line, an online store, or a field, you can find us busy at work, hands in the clay - at least at our office in Helsinki.</em></p>]]></content:encoded></item><item><title><![CDATA[AI ethics guidelines and their shortcomings]]></title><description><![CDATA[Many organizations have created and published their own AI ethics guidelines to promote responsible and ethical development and use of AI. This blog post shows that having principled AI ethics guidelines alone is not enough to guarantee ethical conduct.]]></description><link>https://blog.emblica.fi/ai-ethics-guidelines-and-their-shortcomings/</link><guid isPermaLink="false">63be693e0f949e0001e2ed5b</guid><dc:creator><![CDATA[Rick Joosten]]></dc:creator><pubDate>Tue, 17 Jan 2023 08:00:24 GMT</pubDate><media:content url="https://blog.emblica.fi/content/images/2023/01/a_background_banner_image_representing_the_ethical_landscape_of_AI_0.png" medium="image"/><content:encoded><![CDATA[<img src="https://blog.emblica.fi/content/images/2023/01/a_background_banner_image_representing_the_ethical_landscape_of_AI_0.png" alt="AI ethics guidelines and their shortcomings"><p><em>Lyhyesti suomeksi: Monet organisaatiot ovat luoneet ja julkaisseet omia teko&#xE4;lyn eettisi&#xE4; ohjeistuksia edist&#xE4;&#xE4;kseen teko&#xE4;lyn vastuullista ja eettist&#xE4; kehitt&#xE4;mist&#xE4; ja k&#xE4;ytt&#xF6;&#xE4;. T&#xE4;m&#xE4; blogikirjoitus kertoo, miksi periaatteelliset teko&#xE4;lyn eettiset ohjeet eiv&#xE4;t yksin riit&#xE4; takaamaan eettisyytt&#xE4;.</em></p><p>The ethics of artificial intelligence (AI) is a broad research area that exists between philosophy and computer science. It encompasses everything from AI accountability (questions such as &quot;who is responsible when an AI system makes an error&quot;) to fairness and transparency. Many companies, organizations, governments, and institutions have created and published their own AI ethics principles and guidelines to promote responsible and ethical development and use of AI. </p><p>This blog will discuss what these principles and guidelines look like and whether self-created guidelines are actually useful when it comes to doing business ethically. This discussion will mostly be based on Brent Mittelstadt&#x2019;s <a href="https://arxiv.org/abs/1906.06668">paper</a> <em>Principles alone cannot guarantee ethical AI</em> of which there will be a short summary, and other sources and references are linked to in the text. Finally, there is a brief section describing ways we can improve or supplement AI ethics guidelines.</p><h2 id="ai-ethics-statements">AI ethics statements</h2><p>Discussing the ethics of AI technology initially was the domain of academics. Early philosophical discussions on artificial general intelligence and machine consciousness made way for discussions about how we can ensure the ethical use of AI systems as AI technology advanced. Most of the current output from research, private, and political organizations are in the form of ethics guidelines such as the <a href="https://digital-strategy.ec.europa.eu/en/library/ethics-guidelines-trustworthy-ai">EU&#x2019;s high-level expert group on AI&#x2019;s guidelines</a>. These guidelines are designed to be used as a framework to give guidance on how to deal with ethical issues and inform policy decisions but are not legally binding in and of themselves. </p><p>The German organization AlgorithmWatch started collecting and labeling these AI ethics guidelines in 2019. Currently, their inventory holds 167 guidelines from various companies, organizations, and governments (this does not include legislation). These guidelines are mostly focused on systems for automated decision-making. Of these guidelines, only 8 have a binding agreement describing an enforcement mechanism. All others only contain either voluntary commitments or simple recommendations and principles. Furthermore, the guidelines come majorly from Europe and the US. This shows a heavy influence of western values on publicly available <a href="https://arxiv.org/abs/1906.11668">guidelines</a> with almost no representation of African, South and Central American, and Central Asian countries (excluding India).<br></p><p>Current ethics initiatives started by the industry can be seen as virtue-signaling according to <a href="https://arxiv.org/abs/1906.11668">critics</a> <a href="https://link.springer.com/article/10.1007/s43681-020-00008-1">&#xB2;</a> <a href="https://link.springer.com/article/10.1007/s11023-020-09517-8">&#xB3;</a> <a href="https://www.emerald.com/insight/content/doi/10.1108/JICES-12-2019-0138/full/html">&#x2074;</a> <a href="https://link.springer.com/article/10.1007/s11023-018-9482-5">&#x2075;</a>. AI development companies have a vested interest to appear to be self-regulating in order to delay any legislation that would make their work more difficult to do or that would cut into their profits. So by publishing their own guidelines, they can present themselves as responsible actors in AI development. Whether this is the intent of the companies or not is up for debate, but by publishing guidelines companies can point towards it and claim to conduct business ethically even when the guidelines are limited in scope. The main limitation is often that the guidelines stick with high-level principles. For example, <a href="https://ai.google/responsibilities/responsible-ai-practices/">Google&#x2019;s</a> <em>recommended practices for AI </em>(one of the shortest in the industry)<em> </em>contains: &#x201C;Use a human-centered design approach&#x201D;. This is expanded by some advice on what a <em>human-centered design approach</em> such as <em>&#x201C;Design features with appropriate disclosures built-in&#x201D; </em>without mentioning what appropriate disclosures are. Another noteworthy limitation is that only a few <a href="https://link.springer.com/article/10.1007/s11023-020-09517-8">guidelines</a> explicitly address topics like the possible effects of AI systems on democratic control, political abuse of AI systems, or the ability of AI systems to reduce social cohesion by creating echo chambers. The only ones that do are created by research institutes, the <a href="https://www.montrealdeclaration-responsibleai.com/">Montr&#xE9;al Declaration for Responsible Development of Artificial Intelligence</a>, and the <a href="https://ainowinstitute.org/AI_Now_2019_Report.pdf">AI Now 2019 Report</a>. The latter report is also the <a href="https://link.springer.com/article/10.1007/s11023-020-09517-8">only one </a>discussing &#x201C;hidden&#x201D; social and ecological costs such as AI technologies creating consumer practices that contradict sustainability goals. To meet sustainability goals, care should be taken to reflect on the need to mine one-way-use materials, the energy consumption of running AI services, or how electronic waste should be handled.</p><p>Critics say that most corporate AI ethics discussions should be described as ethics-washing. Ethics-washing can refer to cases where there is no real mechanism to implement these guidelines or to assess that development is more ethically aware. On the other hand, reducing ethical questions to solvable technological problems will also result in ethics-washing. While creating technical solutions to identify and mitigate ethical problems can be necessary to turn a high-level principle into practice, seeing most ethical problems as solvable using technology overlooks the broader social problems and reduces continuous discussion. Later in this post, there will be more discussion on how to make sure AI ethics is not simply &#x201C;technological solutionism&#x201D;.</p><p>While principles give broad gestures where it is easy to find consensus among large and diverse groups of people, they do not give any moralistic judgment. For example, the principle of &#x201C;AI should be fair&#x201D; is widely accepted but does not indicate what is considered fair. In some situations, equality, where everyone is treated the exact same, is preferred. In other, you might want to give some preferential treatment to compensate for a group that has historically been disadvantaged. It is up to the individuals who are implementing some AI system to make a decision about what fairness means in this case. The ethics principles only dictate that it should be fair, but the interpretation of how it should be operationalized is up to individuals. This shows that ethical principles can help in guiding ethical AI development but cannot guarantee the final outcomes of an AI system developed with these principles provides an ethical outcome. Most AI ethics researchers seek to complement or change the principled approach to dealing with AI ethics. There are <a href="https://www.emerald.com/insight/content/doi/10.1108/JICES-12-2019-0138/full/html#sec77">some</a>, however, that claim that if the principles are comprehensive enough, the guidelines will not oversimplify the ethical debate.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://blog.emblica.fi/content/images/2023/01/Drawing-of-fairness-and-equality-in-an-ethical-landscape-1.jpg" class="kg-image" alt="AI ethics guidelines and their shortcomings" loading="lazy" width="512" height="384"><figcaption>Drawing of fairness and equality in an ethical landscape via Stable Diffusion.</figcaption></figure><h2 id="principles-cannot-guarantee-ethical-ai">Principles cannot guarantee ethical AI</h2><p>In the paper <a href="https://arxiv.org/abs/1906.06668"><em>Principles alone cannot guarantee ethical AI</em></a><em>, </em>Brent Mittelstadt takes the position that guidelines with principles are not enough to ensure AI organizations will behave ethically. The stance is reinforced by <a href="https://dl.acm.org/doi/10.1145/3236024.3264833">empirical experiments</a> showing that reading ethics guidelines have no significant influence on the ethical decision-making of software developers. When looking at publicly available ethics guidelines, <a href="https://link.springer.com/article/10.1007/s11023-018-9482-5">reviews</a> <a href="https://arxiv.org/abs/1906.11668">show</a> that most center around four main principles, respect for human autonomy, prevention of harm, fairness, and explainability. These principles are similar to medical ethics which have a long history of principled ethical standards. However, Mittelstadt points out four main differences between medical ethics and AI ethics and critically examines why ethical principles are suitable for the field of medicine, but not for AI development. This section will discuss these differences and go over suggestions on how we can supplement these ethical principles.</p><h3 id="differences-between-principled-medical-ethics-and-ai-ethics">Differences between principled medical ethics and AI ethics</h3><h4 id="1-common-aims-and-fiduciary-duties">1 Common aims and fiduciary duties</h4><p>A medical professional has a moral duty to look after the well-being of their patient. This primary goal is absent in AI development. AI developers do not have something that could be equivalent to a &#x201C;patient&#x201D;. Due to the lack of a fiduciary relationship with a &#x201C;patient&#x201D; and common pressures to behave against the public interest, users of AI systems cannot rely on the developers to keep their best interests in mind.</p><h4 id="2-professional-history-and-norms">2 Professional history and norms</h4><p>Medical professional history is significantly longer than that of AI development going back as far as the Hippocratic oath in the western world. Throughout history, ethical guidelines have been created to serve as a basis for &#x2018;good&#x2019; clinical decision-making. These long-standing traditions can be referenced when having to translate and contextualize a high-level ethical guideline into practice. AI development cannot reference long-standing norms. Furthermore, AI development varies wildly in approach, goals, backgrounds, histories, and moral obligations. These two things combined leaves the developers to interpret and resolve ethical dilemmas as they see fit. Without a large history and common goal to guide them.</p><h4 id="3-proven-methods-to-translate-principles-into-practice">3 Proven methods to translate principles into practice</h4><p>High-level principles do not automatically translate into practice. A high-level principle has to be translated into norms and then into practical requirements. The normative decisions made while translating high-level principles must take into account the specific technology, applications, and local norms. This will lead to conflicting requirements depending on the context in which the principles were translated. Because of the need to interpret the principles depending on the context and the lack of longstanding historical norms, principles do not guarantee consistent ethical behavior. This could be alleviated by creating sector- and case-specific guidelines, technical solutions, and empirical knowledge of how AI can impact its surroundings instead of top-down general guidelines.</p><p>Compared to the field of medicine, there are also no common bodies that help determine what is ethical behavior in day-to-day practice or assessing particularly difficult cases. Bodies like these create proven methods to translate principles into practice across the industry.</p><h4 id="4-legal-and-professional-accountability-mechanisms">4 Legal and professional accountability mechanisms</h4><p>Lastly, AI developers do not have any legal and professional accountability mechanisms. A doctor who has behaved unethically can be barred from practicing medicine again because there are licensing schemes as well as ethics committees and professional medical boards. These mechanisms do not exist universally across all of AI development. Because of their absence, an AI developer cannot be sanctioned for not following ethical principles so the user cannot take it for granted that the AI developer has committed to some ethical framework. Additionally, there is no set pathway to correct wrongdoings if they occur. </p><p>As mentioned in the previous point, incorporating independent ethical auditing into AI development can keep bad actors responsible as well as assess difficult ethical cases, and set a precedent when they come up. Mittelstadt is by far not the only person to advocate for independent ethics boards. Most notably, the AI Now Report [<a href="https://ainowinstitute.org/AI_Now_2018_Report.pdf">2018</a>, <a href="https://ainowinstitute.org/AI_Now_2019_Report.pdf">2019</a>] keeps repeating that the AI industry needs to have new approaches to governance in order to ensure ethical AI use. There are already initiatives for creating new governance mechanisms such as <a href="https://ai-governance.eu/">AIGA</a> here in Finland. Moreover, with an outside organization, it would be possible to license developers who work on high-risk AI.</p><h3 id="conclusion">Conclusion</h3><p>This blog showed the shortcomings of having high-level ethical principles. Many solutions to overcome these shortcomings require a shift in AI development by increasing outside oversight and legislation to ensure ethical business. While individual organizations cannot achieve this alone, they should be aware of the traps and weaknesses that high-level ethics principles bring with them. Most importantly, ethics will always be a process. We shouldn&#x2019;t see ethical challenges as something that can be fixed through technical solutions and principles alone.</p><p>The constant changes in technology and applications require us to approach ethics as an ever-changing process. New ethical challenges will come up and require an open debate in order to make sure AI development reflects many viewpoints and ethical questions are not simplified to &#x2018;solvable&#x2019; problems. A first step towards continued ethical thinking could be incorporating AI ethics checklists into AI development. These checklists can guide developers to embrace ethical thinking as long as the checklist is broad enough as well as require active participation from the <a href="https://www.emerald.com/insight/content/doi/10.1108/JICES-12-2019-0138/full/html#sec77">participants</a>.</p><h2 id="emblica-and-ethics">Emblica and ethics</h2><p>Here at Emblica, we are strengthening our ethical commitments by implementing an active ethics review into our workflow. We have been reviewing statements and guidelines from several companies and institutions to gauge what is out there and what we can learn from them. As this blog post discussed, only having a set of AI principles cannot guarantee that our processes are ethical. That is why we are taking our time to review and consider what approaches will give the best practical ethics recommendations and also avoid the trap of ethics-washing. Once we have concluded our research we will publish our approach to ensure that our AI solutions are developed ethically.</p><hr><p><a href="https://blog.emblica.fi/ghost/emblica.com"><em>Emblica</em></a><em> is a technology company focused on data-intensive applications and artificial intelligence. Some examples of our customers are Sanoma, Uponor, Caruna, and the Tax Administration. Emblica is 100% owned by its employees.</em></p>]]></content:encoded></item></channel></rss>