<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Data Processing Archives - OVHcloud Blog</title>
	<atom:link href="https://blog.ovhcloud.com/tag/data-processing/feed/" rel="self" type="application/rss+xml" />
	<link>https://blog.ovhcloud.com/tag/data-processing/</link>
	<description>Innovation for Freedom</description>
	<lastBuildDate>Mon, 03 Jul 2023 08:46:27 +0000</lastBuildDate>
	<language>en-GB</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>

<image>
	<url>https://blog.ovhcloud.com/wp-content/uploads/2019/07/cropped-cropped-nouveau-logo-ovh-rebranding-32x32.gif</url>
	<title>Data Processing Archives - OVHcloud Blog</title>
	<link>https://blog.ovhcloud.com/tag/data-processing/</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>How to tackle data problems</title>
		<link>https://blog.ovhcloud.com/how-to-tackle-data-problems/</link>
		
		<dc:creator><![CDATA[Philip Marais]]></dc:creator>
		<pubDate>Thu, 17 Feb 2022 16:29:14 +0000</pubDate>
				<category><![CDATA[OVHcloud Startup Program]]></category>
		<category><![CDATA[Data Platform]]></category>
		<category><![CDATA[Data Processing]]></category>
		<category><![CDATA[Startup Program]]></category>
		<guid isPermaLink="false">https://blog.ovhcloud.com/?p=22478</guid>

					<description><![CDATA[Let&#8217;s say you have a data problem&#8230; Maybe you want to extract some value from your data, for example to give your app an advantage in the market. You look at all the data you have stored, and at the moment it looks like the inside of your house during lockdown&#8230; You did a search [&#8230;]<img src="//blog.ovhcloud.com/wp-content/plugins/matomo/app/matomo.php?idsite=1&amp;rec=1&amp;url=https%3A%2F%2Fblog.ovhcloud.com%2Fhow-to-tackle-data-problems%2F&amp;action_name=How%20to%20tackle%20data%20problems&amp;urlref=https%3A%2F%2Fblog.ovhcloud.com%2Ffeed%2F" style="border:0;width:0;height:0" width="0" height="0" alt="" />]]></description>
										<content:encoded><![CDATA[
<h3 class="wp-block-heading" id="let-s-say-you-have-a-data-problem">Let&#8217;s say you have a data problem&#8230; </h3>



<p>Maybe you want to extract some value from your data, for example to give your app an advantage in the market. You look at all the data you have stored, and at the moment it looks like the inside of your house during lockdown&#8230; You did a search on what tools to use and you ended up on this <a href="http://mattturck.com/wp-content/uploads/2019/07/2019_Matt_Turck_Big_Data_Landscape_Final_Fullsize.png" target="_blank" rel="noreferrer noopener nofollow external" data-wpel-link="external">page</a>, so you got more questions than answers. As CFO you can speak some of the data lingo but you are not going to be mastering advanced statistics or machine learning models any time soon.</p>



<div class="wp-block-image"><figure class="aligncenter size-large is-resized"><img fetchpriority="high" decoding="async" src="https://blog.ovhcloud.com/wp-content/uploads/2022/02/IMG_0816-1024x537.jpeg" alt="How to tackle data problems" class="wp-image-22488" width="512" height="269" srcset="https://blog.ovhcloud.com/wp-content/uploads/2022/02/IMG_0816-1024x537.jpeg 1024w, https://blog.ovhcloud.com/wp-content/uploads/2022/02/IMG_0816-300x157.jpeg 300w, https://blog.ovhcloud.com/wp-content/uploads/2022/02/IMG_0816-768x403.jpeg 768w, https://blog.ovhcloud.com/wp-content/uploads/2022/02/IMG_0816.jpeg 1200w" sizes="(max-width: 512px) 100vw, 512px" /></figure></div>



<h3 class="wp-block-heading" id="so-what-do-you-do">So, what do you do?</h3>



<p> Well, the first step is to <strong>make sure you understand what the problem</strong> is you are trying to solve. Is it an AI problem or could it be solved with a simpler analytics tool? Presumably you have a business problem, and you hope that you will find the answer in your chaotic data. You need to phrase this business problem as a data problem in order to<strong> identify what tool to use</strong> to extract the answer. </p>



<div class="wp-block-image"><figure class="aligncenter size-large"><img decoding="async" width="1024" height="634" src="https://blog.ovhcloud.com/wp-content/uploads/2022/02/IMG_0815-1024x634.jpeg" alt=" Well, the first step is to make sure you understand what the problem is you are trying to solve. Is it an AI problem or could it be solved with a simpler analytics tool? Presumably you have a business problem, and you hope that you will find the answer in your chaotic data. You need to phrase this business problem as a data problem in order to identify what tool to use to extract the answer. " class="wp-image-22486" srcset="https://blog.ovhcloud.com/wp-content/uploads/2022/02/IMG_0815-1024x634.jpeg 1024w, https://blog.ovhcloud.com/wp-content/uploads/2022/02/IMG_0815-300x186.jpeg 300w, https://blog.ovhcloud.com/wp-content/uploads/2022/02/IMG_0815-768x476.jpeg 768w, https://blog.ovhcloud.com/wp-content/uploads/2022/02/IMG_0815.jpeg 1536w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure></div>



<p>Do you need a machine to<strong> solve your problem like a human</strong> would? Then you have a problem that <strong>artificial intelligence (AI</strong>) could assist with. Or do you need an algorithm that<strong> learns from examples</strong> without being explicitly programmed? Then you have a problem that <strong>machine learning (ML)</strong> could assist with. Or do you need to <strong>deliver some pretty graphs</strong> once you have cleaned up your data? Then you have a problem that <strong>business intelligence (BI) </strong>tools could assist with. Or can the solution be delivered by<strong> programming a few rules</strong>? Then you have a problem that could be solved with <strong>conventional programming</strong>. These are some of the questions you need to answer.</p>



<p>If you just need to try and<strong> determine what happened </strong>from your data (i.e., be descriptive) you may be able to do this with a <strong>BI tool or conventional programming</strong>. You may need a <strong>machine learning</strong> solution if you need to determine what will happen (i.e.,<strong> obtain a prediction</strong>) or what to do (i.e., <strong>provide prescriptions</strong>) from your data. Typical predictive questions may be “what music will my customer want to download through my app next month?” or “how much discount should I provide to my customer next week?”. Typical prescriptive questions would be “how should we segment our customers?” or “what topics are in our customer feedback?”.</p>



<h3 class="wp-block-heading" id="aligning-objectives-and-expectations">Aligning objectives and expectations</h3>



<p>Of course, the defined problem needs to<strong> align with your business objectives and expectations</strong>. What are your strategic objectives, what are the timeframes for delivery and are the risks understood? Do you have the necessary skills and resources for implementation? Lack of alignment could hamper reaching the objective further down the process.</p>



<p>The above questions will all help you to<strong> translate your business objectives into data related objectives</strong> and goals and <strong>define your problem statement</strong>. Your problem statement will define the problem, the impact it has and what has been identified as the best starting point to solve the problem. The starting point will include one or more of your data related goals. </p>



<p>Remember that your <strong>objective needs to be specific</strong>. An objective like, “We want to use AI to make our product better that the competition”, does not really communicate what problem you are trying to solve and is unlikely to align with business objectives. An objective like, “We want to predict our customers’ next product choices in order to give them a better customer experience” helps to clarify the business problem and the data problem.</p>



<h3 class="wp-block-heading" id="choosing-the-right-technology">Choosing the right technology</h3>



<p>Once the above questions have been answered, the objectives are clear and when the problem is well defined it is then <strong>much easier to choose the correct technology</strong> to address the problem. From there you can dive deeper into the specific data requirements and resource needs to use that technology. If you are interested to find out more about mastering your data, we are running an <strong>AI and data focused event on 10 March 2022</strong> that includes a roundtable on this topic and more. The key takeaways will be:</p>



<p>Find out more about the event and<strong> sign up <a href="https://hopin.com/events/super-connect-series-showcase-ai/registration" data-wpel-link="external" target="_blank" rel="nofollow external noopener noreferrer">here</a></strong> or you can find out more about <strong>our managed cloud solutions <a href="https://www.ovhcloud.com/en-gb/public-cloud/" data-wpel-link="external" target="_blank" rel="nofollow external noopener noreferrer">here</a>.</strong></p>



<div class="wp-block-image"><figure class="aligncenter size-full is-resized"><a href="https://startup.ovhcloud.com/" data-wpel-link="external" target="_blank" rel="nofollow external noopener noreferrer"><img decoding="async" src="https://blog.ovhcloud.com/wp-content/uploads/2021/06/IMG_0561-05.png" alt="OVHcloud Startup Program" class="wp-image-20943" width="372" height="248" srcset="https://blog.ovhcloud.com/wp-content/uploads/2021/06/IMG_0561-05.png 496w, https://blog.ovhcloud.com/wp-content/uploads/2021/06/IMG_0561-05-300x200.png 300w" sizes="(max-width: 372px) 100vw, 372px" /></a></figure></div>



<p>To find out more about the OVHcloud Startup Program, or to sign up for free credits and technical support, go to <a href="https://startup.ovhcloud.com/" data-wpel-link="external" target="_blank" rel="nofollow external noopener noreferrer">https://startup.ovhcloud.com/</a></p>
<img loading="lazy" decoding="async" src="//blog.ovhcloud.com/wp-content/plugins/matomo/app/matomo.php?idsite=1&amp;rec=1&amp;url=https%3A%2F%2Fblog.ovhcloud.com%2Fhow-to-tackle-data-problems%2F&amp;action_name=How%20to%20tackle%20data%20problems&amp;urlref=https%3A%2F%2Fblog.ovhcloud.com%2Ffeed%2F" style="border:0;width:0;height:0" width="0" height="0" alt="" />]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The new Logs Data Platform</title>
		<link>https://blog.ovhcloud.com/the-new-logs-data-platform/</link>
		
		<dc:creator><![CDATA[Carington Lucas Saint George]]></dc:creator>
		<pubDate>Mon, 26 Apr 2021 11:12:00 +0000</pubDate>
				<category><![CDATA[OVHcloud Product News]]></category>
		<category><![CDATA[Data Platform]]></category>
		<category><![CDATA[Data Processing]]></category>
		<category><![CDATA[Logs Data Platform]]></category>
		<guid isPermaLink="false">https://www.ovh.com/blog/?p=20753</guid>

					<description><![CDATA[At OVHcloud, we see many different use cases around log management. In addition, we have recently had the opportunity to discuss with several client companies with different approaches and maturity on these topics. Based on these insights, we have made major changes to our LDP solution to address these new ways of consuming logs. In [&#8230;]<img src="//blog.ovhcloud.com/wp-content/plugins/matomo/app/matomo.php?idsite=1&amp;rec=1&amp;url=https%3A%2F%2Fblog.ovhcloud.com%2Fthe-new-logs-data-platform%2F&amp;action_name=The%20new%20Logs%20Data%20Platform&amp;urlref=https%3A%2F%2Fblog.ovhcloud.com%2Ffeed%2F" style="border:0;width:0;height:0" width="0" height="0" alt="" />]]></description>
										<content:encoded><![CDATA[
<p>At OVHcloud, we see many different use cases around log management. In addition, we have recently had the opportunity to discuss with several client companies with different approaches and maturity on these topics. Based on these insights, we have made major changes to our LDP solution to address these new ways of consuming logs.</p>



<div class="wp-block-image"><figure class="aligncenter size-large is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2021/04/IMG_0534-1024x538.png" alt="The new Logs Data Platform" class="wp-image-20793" width="768" height="404" srcset="https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0534-1024x538.png 1024w, https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0534-300x158.png 300w, https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0534-768x403.png 768w, https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0534.png 1198w" sizes="auto, (max-width: 768px) 100vw, 768px" /></figure></div>



<p>In this blog post we will see the historical features that made the success of this platform and the improvements resulting from your feedback.</p>



<h3 class="wp-block-heading">Let&#8217;s start from the beginning</h3>



<h4 class="wp-block-heading">Why are logs important?</h4>



<p>It’s critical you know at anytime, what’s going on into your Information System and applications.. You must be able to understand, analyze and monitor the health of your IT and applications to fix any issue. Log files are created each time an event occurs on your computer systems. They therefore provide useful information about the state of your services and infrastructure. These files come from different sources (physical servers, virtual servers, mobile applications, websites, etc.) and are counted in the millions every day, making them difficult to analyze. From your logs, you can extract valuable information about the behavior of your customers or your systems, and then act accordingly.</p>



<div class="wp-block-image"><figure class="aligncenter size-large"><img loading="lazy" decoding="async" width="768" height="455" src="https://www.ovh.com/blog/wp-content/uploads/2021/03/image2021-1-26_2-48-1.png" alt="" class="wp-image-20754" srcset="https://blog.ovhcloud.com/wp-content/uploads/2021/03/image2021-1-26_2-48-1.png 768w, https://blog.ovhcloud.com/wp-content/uploads/2021/03/image2021-1-26_2-48-1-300x178.png 300w" sizes="auto, (max-width: 768px) 100vw, 768px" /></figure></div>



<h4 class="wp-block-heading">How do I analyze my logs?</h4>



<p>With so many logs to collect, store and analyse, we know most IT administrators need turnkey integrated solution. That what we had in mind when we designed our platform. Indeed, Logs Data Platform is a turnkey solution, which allows you to collect, store and analyze logs. It supports the different log files, whether they are related to applications, servers or security. You have the possibility to use the log collector of your choice (Syslog-ng, Fluentd, NXLog) or to use our dedicated collectors (Flowgger and Logstash). These collectors work regardless of the source, format or structure of your data. For analysis, you can rely on visualization via Graylog, Kibana or Grafana. For example, in the context of infrastructure supervision, you can monitor logs at the application or server level. By making you benefit from the ELK (Elasticsearch Logstash Kibana) ecosystem, Logs Data Platform is a powerful log analysis solution.</p>



<div class="wp-block-image"><figure class="aligncenter size-large is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2021/04/IMG_0533-1024x418.png" alt="Logs Data Platform" class="wp-image-20791" width="1024" height="418" srcset="https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0533-1024x418.png 1024w, https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0533-300x123.png 300w, https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0533-768x314.png 768w, https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0533-1536x628.png 1536w, https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0533.png 1799w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure></div>



<div style="height:100px" aria-hidden="true" class="wp-block-spacer"></div>



<h2 class="wp-block-heading"><strong>3 ways to use Logs Data Platform</strong></h2>



<p>We saw that we could diffrenciate 3 major way to store and analyze logs, depending on the business usecase. So we decided that for each log stream you collect, you can activate any of the 3 following approach (or all 3 at the same time) :</p>



<div class="wp-block-image is-style-default"><figure class="aligncenter size-full is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2021/03/21.png" alt="" class="wp-image-20757" width="658" height="413" srcset="https://blog.ovhcloud.com/wp-content/uploads/2021/03/21.png 877w, https://blog.ovhcloud.com/wp-content/uploads/2021/03/21-300x188.png 300w, https://blog.ovhcloud.com/wp-content/uploads/2021/03/21-768x483.png 768w" sizes="auto, (max-width: 658px) 100vw, 658px" /></figure></div>



<div style="height:100px" aria-hidden="true" class="wp-block-spacer"></div>



<div class="wp-block-group"><div class="wp-block-group__inner-container is-layout-flow wp-block-group-is-layout-flow">
<ul class="wp-block-list"><li><strong>WebSocket broadcasting</strong> allows you to see what&#8217;s going on in your application or server in real time. Indeed with this feature our Logs Data Platform allows you to connect different applications or servers to one unique endpoint and make all of them appear in one stream if needed.&nbsp;<strong>ldp-tail</strong>&nbsp;is able to follow one your stream in real-time with sub-second latency.</li></ul>



<p><a href="https://docs.ovh.com/fr/logs-data-platform/ldp-tail/" data-wpel-link="exclude">Follow the guide!</a></p>
</div></div>



<div class="wp-block-group"><div class="wp-block-group__inner-container is-layout-flow wp-block-group-is-layout-flow">
<div class="wp-block-image"><figure class="aligncenter size-large is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2021/03/final.gif" alt="" class="wp-image-20758" width="762" height="410"/></figure></div>
</div></div>



<ul class="wp-block-list"><li>Logs Data Platform allows you to <strong>index</strong> your logs with a flexible retention period ranging from 14 days to 1 year, which allows you to analyze the data over a given period of time.</li></ul>



<p><a href="https://docs.ovh.com/fr/logs-data-platform/quick-start/" data-wpel-link="exclude">Follow the guide!</a></p>



<div class="wp-block-image"><figure class="aligncenter size-large is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2021/03/3.png" alt="" class="wp-image-20759" width="650" height="284" srcset="https://blog.ovhcloud.com/wp-content/uploads/2021/03/3.png 866w, https://blog.ovhcloud.com/wp-content/uploads/2021/03/3-300x131.png 300w, https://blog.ovhcloud.com/wp-content/uploads/2021/03/3-768x336.png 768w" sizes="auto, (max-width: 650px) 100vw, 650px" /></figure></div>



<ul class="wp-block-list"><li>You also have the possibility to archive your logs for a long period of time (from 1 year to 10 years) thanks to the <strong>Cold Storage</strong> feature of Logs Data Platform. This can be very useful for example within the framework of the GDPR requirements or simply to keep the log history of your infrastructure.</li></ul>



<p><a href="https://docs.ovh.com/fr/logs-data-platform/cold-storage/" data-wpel-link="exclude">Follow the guide!</a></p>



<div class="wp-block-image"><figure class="aligncenter size-large is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2021/03/4.png" alt="" class="wp-image-20760" width="396" height="362" srcset="https://blog.ovhcloud.com/wp-content/uploads/2021/03/4.png 528w, https://blog.ovhcloud.com/wp-content/uploads/2021/03/4-300x274.png 300w" sizes="auto, (max-width: 396px) 100vw, 396px" /></figure></div>



<h3 class="wp-block-heading">How does it work?</h3>



<p>Now, I think you see broadly how you can leverage the platform. Let&#8217;s dig deeper on the technologies that power it and how you can leverage them for ingest, query and analysis.</p>



<h4 class="wp-block-heading" id="Blogpost(draft)-Ingestion">Ingestion</h4>



<p>Logs Data Platform is compatible with most of the standard protocols on the market: GELF, SYSLOG, Cap&#8217;n&#8217;Proto, LTSV, RFC5425, Beats&#8230;</p>



<p>Moreover you also have the possibility to subscribe to dedicated collectors such as Logstash or Flowgger for more flexibility.</p>



<h4 class="wp-block-heading" id="Blogpost(draft)-Query">Query</h4>



<p>If you have chosen to index your logs, then you have different ways to analyze the results: you can choose to use one of the visualization tools provided by OVHcloud (Graylog, Grafana or Kibana) or use the Elasticsearch, Graylog or OVHcloud APIs in order to use your own analysis tools.</p>



<div class="wp-block-image"><figure class="aligncenter size-large"><img loading="lazy" decoding="async" width="1024" height="690" src="https://www.ovh.com/blog/wp-content/uploads/2021/04/IMG_0542-1024x690.png" alt="Logs Data Platform" class="wp-image-20807" srcset="https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0542-1024x690.png 1024w, https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0542-300x202.png 300w, https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0542-768x517.png 768w, https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0542-1536x1034.png 1536w, https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0542-2048x1379.png 2048w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure></div>



<div style="height:100px" aria-hidden="true" class="wp-block-spacer"></div>



<h3 class="wp-block-heading">And the little bonus</h3>



<p>Logs Data Platform also allows you to index data other than logs thanks to its <strong>Index as a Service</strong> feature based on Elasticsearch, you can for example index documents.<br>Thanks to this feature you can for example create powerful search engines thanks to the performance of Elasticsearch and all this without worrying about the integration of Elasticsearch because the Index as a Service of Logs Data Platform is a turnkey solution fully managed by OVHcloud.</p>



<p><a href="https://docs.ovh.com/fr/logs-data-platform/index-as-a-service/" data-wpel-link="exclude">Follow the guide!</a></p>



<div class="wp-block-image"><figure class="aligncenter size-large is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2021/04/IMG_0537.png" alt="" class="wp-image-20801" width="313" height="438" srcset="https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0537.png 625w, https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0537-214x300.png 214w" sizes="auto, (max-width: 313px) 100vw, 313px" /></figure></div>



<h3 class="wp-block-heading">So what&#8217;s new in this new version?</h3>



<div class="wp-block-image"><figure class="alignright size-large is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2021/04/IMG_0536.png" alt="" class="wp-image-20799" width="341" height="256" srcset="https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0536.png 682w, https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0536-300x225.png 300w" sizes="auto, (max-width: 341px) 100vw, 341px" /></figure></div>



<ul class="wp-block-list"><li>You expressed us the wish to have a more flexible invoicing, so we changed our pricing model to pay-as-you-go. Indeed, pay-as-you-go makes invoicing simpler, more readable and predictable. Moreover, you can now take advantage of thresholds and alerts to improve your consumption efficiency.</li></ul>



<ul class="wp-block-list"><li>Until recently our logs analysis platform was only available for French customers, now Logs Data Platform is available in all countries and in all languages!</li></ul>



<div class="wp-block-image"><figure class="aligncenter size-large"><img loading="lazy" decoding="async" width="352" height="364" src="https://www.ovh.com/blog/wp-content/uploads/2021/03/image2021-1-26_5-22-45.png" alt="" class="wp-image-20764" srcset="https://blog.ovhcloud.com/wp-content/uploads/2021/03/image2021-1-26_5-22-45.png 352w, https://blog.ovhcloud.com/wp-content/uploads/2021/03/image2021-1-26_5-22-45-290x300.png 290w" sizes="auto, (max-width: 352px) 100vw, 352px" /></figure></div>



<div class="wp-block-image"><figure class="alignright size-large is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2021/04/IMG_0535.png" alt="" class="wp-image-20798" width="273" height="233" srcset="https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0535.png 545w, https://blog.ovhcloud.com/wp-content/uploads/2021/04/IMG_0535-300x257.png 300w" sizes="auto, (max-width: 273px) 100vw, 273px" /></figure></div>



<ul class="wp-block-list"><li>Security and confidentiality are becoming more and more important in company policies, so in order to comply with your stringent security requirements, we have created the Enterprise Logs Account on Logs Data Platform. Thanks to this dedicated cluster you are totally isolated. It will allow us to offer you brand new features such as the Network Access Control List or customizable retention period.</li></ul>



<h3 class="wp-block-heading">Moreover</h3>



<p>Moreover, Logs Data Platform will soon be ISO/IEC certified.<br>Indeed, we are only a few weeks away from obtaining the ISO/IEC 27001, 27017, 27018 and 27701 certificates.<br>What do these norms correspond to?</p>



<p>We won&#8217;t go into boring legal details here, but to put it simply:</p>



<ul class="wp-block-list"><li>ISO/IEC 27001:2013 Certification and ISMS relating to Information security management systems for cloud services</li><li>ISO/IEC 27017:2015 Certification relating to information security controls for cloud services</li><li>ISO/IEC 27018:2014 Code of practice for protection of personally identifiable information for cloud services</li><li>ISO/IEC 27701:2019 Certification and PIMS relating to personal data processing security management</li></ul>



<p>To summarize, these ISO/IEC certifications ensure the presence of an Information Security Management System (ISMS) for the management of risks, vulnerabilities and business continuity, as well as a Privacy Information Management System (PIMS).</p>



<p>That said, if the legal is your passion, you&#8217;ll find more details on <a href="http://iso.org" data-wpel-link="external" target="_blank" rel="nofollow external noopener noreferrer">iso.org</a>.</p>



<p>And that&#8217;s not all, at the same time our Enterprise Logs Account offer will be HDS compatible to host health data. You will find more information <a href="https://www.ovhcloud.com/en-ie/enterprise/certification-conformity/hds/" data-wpel-link="external" target="_blank" rel="nofollow external noopener noreferrer">here</a>.</p>



<hr class="wp-block-separator"/>



<h3 class="wp-block-heading">NB: </h3>



<p>A few weeks after we released these improvements on our Logs Data Platform Product, Elastic announced changes in licensing for the future versions of Elasticsearch and Kibana offered as a service.</p>



<p>Other members of the Elastic open-source community announced that open versions of those components will continue to exist. Be reassured that the platform as it exists now is not impacted by the change and we will in the mid term future explore the best options to keep offering your the latest feature of the ecosystem, sticking to our S.M.A.R.T. values.</p>



<hr class="wp-block-separator"/>



<h3 class="wp-block-heading">Contact us</h3>



<p><a href="https://community.ovh.com/c/platform/data-platforms" data-wpel-link="exclude">https://community.ovh.com/c/platform/data-platforms</a></p>



<h3 class="wp-block-heading" id="Blogpost(draft)-Someusefullinks">Some useful links</h3>



<ul class="wp-block-list"><li><a href="https://www.ovhcloud.com/en-ie/data-platforms/logs/" data-wpel-link="external" target="_blank" rel="nofollow external noopener noreferrer">https://www.ovhcloud.com/en-ie/data-platforms/logs/</a></li><li><a href="https://docs.ovh.com/fr/logs-data-platform/" data-wpel-link="exclude">https://docs.ovh.com/fr/logs-data-platform/</a></li></ul>
<img loading="lazy" decoding="async" src="//blog.ovhcloud.com/wp-content/plugins/matomo/app/matomo.php?idsite=1&amp;rec=1&amp;url=https%3A%2F%2Fblog.ovhcloud.com%2Fthe-new-logs-data-platform%2F&amp;action_name=The%20new%20Logs%20Data%20Platform&amp;urlref=https%3A%2F%2Fblog.ovhcloud.com%2Ffeed%2F" style="border:0;width:0;height:0" width="0" height="0" alt="" />]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Why are you still managing your data processing clusters?</title>
		<link>https://blog.ovhcloud.com/why-are-you-still-managing-your-data-processing-clusters/</link>
		
		<dc:creator><![CDATA[Mojtaba Imani]]></dc:creator>
		<pubDate>Wed, 30 Sep 2020 16:14:35 +0000</pubDate>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[Apache Spark]]></category>
		<category><![CDATA[Big Data]]></category>
		<category><![CDATA[Cluster Computing]]></category>
		<category><![CDATA[Data Platform]]></category>
		<category><![CDATA[Data Processing]]></category>
		<category><![CDATA[Distributed Computing]]></category>
		<guid isPermaLink="false">https://www.ovh.com/blog/?p=19241</guid>

					<description><![CDATA[Cluster computing is used to share a computation load among a group of computers. This achieves a higher level of performance and scalability.&#160;&#160;&#160; Apache Spark is an open-source, distributed and cluster-computing framework, that is much faster than the previous one (Hadoop MapReduce). This is thanks to features like in-memory processing and lazy evaluation. Apache Spark [&#8230;]<img src="//blog.ovhcloud.com/wp-content/plugins/matomo/app/matomo.php?idsite=1&amp;rec=1&amp;url=https%3A%2F%2Fblog.ovhcloud.com%2Fwhy-are-you-still-managing-your-data-processing-clusters%2F&amp;action_name=Why%20are%20you%20still%20managing%20your%20data%20processing%20clusters%3F&amp;urlref=https%3A%2F%2Fblog.ovhcloud.com%2Ffeed%2F" style="border:0;width:0;height:0" width="0" height="0" alt="" />]]></description>
										<content:encoded><![CDATA[
<p>Cluster computing is used to share a computation load among a group of computers. This achieves a higher level of performance and scalability.&nbsp;&nbsp;&nbsp;</p>



<p>Apache Spark is an open-source, distributed and cluster-computing framework, that is much faster than the previous one (Hadoop MapReduce). This is thanks to features like in-memory processing and lazy evaluation. Apache Spark is the most popular tool in this category.</p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="539" src="https://www.ovh.com/blog/wp-content/uploads/2020/09/IMG_0282-1024x539.png" alt="Why are you still managing your data processing clusters?" class="wp-image-19364" srcset="https://blog.ovhcloud.com/wp-content/uploads/2020/09/IMG_0282-1024x539.png 1024w, https://blog.ovhcloud.com/wp-content/uploads/2020/09/IMG_0282-300x158.png 300w, https://blog.ovhcloud.com/wp-content/uploads/2020/09/IMG_0282-768x404.png 768w, https://blog.ovhcloud.com/wp-content/uploads/2020/09/IMG_0282.png 1200w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p>The analytics engine is the leading framework for large-scale SQL, batch processing, stream processing and machine learning. For coding in Spark, you have the option of using different programming languages; including Java, Scala, Python, R and SQL. It can be run locally on a single machine, or on a cluster of computers for task distribution.</p>



<p>By using Apache Spark, you can process your data in your local computer, or you can create a cluster to send any number of processing jobs.</p>



<p>It is possible to create your cluster with physical computers on-premises, with virtual machines in a hosting company, or with any cloud provider. With your own cluster, you’ll have the ability to send Spark jobs whenever you like.&nbsp;</p>



<h3 class="wp-block-heading">Cluster Management Challenges&nbsp;</h3>



<p>If you are processing a huge amount of data and you expect to have results in a reasonable time, your local computer won’t be enough. You need a cluster of computers to divide the data and process workloads &#8211; several computers are run in parallel to speed up the task.&nbsp;</p>



<p>Creating and managing your own cluster of computers, however, is not an easy task. You will face several challenges:&nbsp;</p>



<h4 class="wp-block-heading">Cluster Creation&nbsp;</h4>



<p>Creating an Apache Spark cluster is an arduous task.&nbsp;</p>



<p>First, you’ll need to create a cluster of computers and install an operating system, development tools (Python, Java, Scala), etc.&nbsp;</p>



<p>Second, you’ll need to select a version of Apache Spark and install the necessary nodes (master&nbsp;and workers).&nbsp;</p>



<p>Lastly, you’ll need to connect all these nodes together to finalize your&nbsp;Apache Spark cluster. </p>



<p>All in all, it can take several hours to create and configure a new Apache Spark cluster.</p>



<h4 class="wp-block-heading">Cluster Management&nbsp;</h4>



<p>But once you have your own cluster up and running, your job is far from over. Is your cluster working well? Is each and every node healthy?&nbsp;</p>



<p>Here is the second challenge: facing the pain of cluster management!</p>



<p>You’ll need to check the health of all your nodes manually or, preferably, install monitoring tools that report any issues nodes may encounter. </p>



<p>Do the nodes have enough disk space available for new tasks? One key issue faced by Apache Spark clusters, is that some tasks write a lot of data in the local disk space of nodes without deleting them. Disk space is a common problem and, as you may know, a lack of disk space eliminates the possibility of running more tasks. &nbsp;</p>



<p>Do you need to run multiple Spark jobs at the same time? Sometimes a single job occupies all the CPU and RAM resources in your cluster and doesn’t allow other jobs to start and run at the same time.&nbsp;</p>



<p>These are only a few of the problems you will meet while working with your own clusters.</p>



<h4 class="wp-block-heading">Cluster Security</h4>



<p>Now for the third challenge! What is even more important than having a cluster up and running smoothly?</p>



<p>You guessed it: security. After all, Apache Spark is a Data Processing tool. And data is very sensitive.</p>



<p>Where in your cluster, does security matter most? </p>



<p>What about the connection between nodes? Are they connected with a secured (and fast) connection? Who has access to the servers housing your cluster?&nbsp;</p>



<p>If you have created your cluster on the cloud and you are working with sensitive data, you’ll need to address these issues by securing each and every node and encrypting communications between them.&nbsp;</p>



<h4 class="wp-block-heading">Spark Version</h4>



<p>Here is your fourth challenge: managing your cluster&#8217;s user expectations. In some cases this may be a less daunting task, but not all. </p>



<p>There isn&#8217;t a whole lot you can do to change the expectations of your cluster&#8217;s users, but here&#8217;s a common example to help you prepare:</p>



<p>Do your users like to test their codes with different versions of Apache Spark?&nbsp; Or do they require the latest feature from the latest Spark nightly version?</p>



<p>When you create an Apache Spark cluster, you have to select one version of Spark. Your whole cluster will be bound to it, and <em>it</em> alone. This means it isn&#8217;t possible for several versions to cohabit in the same cluster.</p>



<p>So, either you’ll have to change the Spark version of your whole cluster or create another separated cluster. And of course, if you decide to do that, you have to create a down time on your cluster to make the modifications.&nbsp;</p>



<h4 class="wp-block-heading">Cluster Efficiency&nbsp;</h4>



<p>And for the final challenge: scaling!</p>



<p>How can you get the most benefit from the cluster resources you are paying for? Are you paying for your cluster but feel you&#8217;re not using it efficiently? Is your cluster too big for your users? Is it running, but empty of jobs during the holiday seasons?</p>



<p>When you have a processing cluster &#8211; especially if you have a lot of precious resources in your cluster that you&#8217;re paying for &#8211; you will always have one major concern: is your cluster being utilised as efficiently as possible. There will be occasions that some resources in your cluster are idle, or where you are only running small jobs that don&#8217;t require the amount of resources in your cluster. Scaling will become a major pain point.</p>



<h3 class="wp-block-heading">OVHcloud Data Processing (ODP) Solution&nbsp;</h3>



<p>At OVHcloud, we created a new data service called OVHcloud Data Processing (ODP) to address all cluster management challenges mentioned above.&nbsp;</p>



<p>Let’s assume that you have some data to process but you don’t have the desire, the time, the budget or the skills to overcome these challenges. Maybe you don’t want to, or can’t, ask for help from colleagues or consultants to spawn and manage a cluster. How can you still make use of Apache Spark? This is where the ODP service comes in!</p>



<p>By using ODP, you need to write your Apache Spark code and ODP will do the rest. It will create a disposable dedicated Apache Spark cluster over the cloud for each job in just a few seconds &#8211; then delete the whole cluster after finishing the job. You only pay for the requested resources and only for the duration of the computation. There is no need to pay for hours and hours of cloud servers, while you are busy with the cluster installation, configuration or even debugging and updating the engine version.&nbsp;</p>



<div class="wp-block-image"><figure class="aligncenter size-large is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2020/07/EF2587E4-9E30-4C3B-A6E1-0EAF3FF8F8F6-1024x813.jpeg" alt="OVHcloud Data Processing" class="wp-image-18856" width="512" height="407" srcset="https://blog.ovhcloud.com/wp-content/uploads/2020/07/EF2587E4-9E30-4C3B-A6E1-0EAF3FF8F8F6-1024x813.jpeg 1024w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/EF2587E4-9E30-4C3B-A6E1-0EAF3FF8F8F6-300x238.jpeg 300w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/EF2587E4-9E30-4C3B-A6E1-0EAF3FF8F8F6-768x610.jpeg 768w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/EF2587E4-9E30-4C3B-A6E1-0EAF3FF8F8F6.jpeg 1227w" sizes="auto, (max-width: 512px) 100vw, 512px" /></figure></div>



<h4 class="wp-block-heading">ODP Cluster Creation</h4>



<p>When you submit your job, ODP will create an apache spark cluster dedicated to that job in just a few seconds. This cluster will have the amount of CPU and RAM and the number of workers specified in the job submit form. All necessary software will be automatically installed. You don’t need to worry at all about a cluster, how to install, configure, or secure it. ODP does all of this for you.&nbsp;</p>



<h4 class="wp-block-heading">ODP Cluster Management&nbsp;</h4>



<p>When you submit your job, cluster management and monitoring are configured and handled by ODP. All logging and monitoring mechanisms and tools will be installed automatically for you. You will have a Grafana dashboard to monitor different parameters and resources of your job and you will have access to the official Apache Spark dashboard.&nbsp;</p>



<p>You don’t need to worry about cleaning the local disk of each node because each job will start with fresh resources. It isn&#8217;t possible, therefore, for one job to delay another job as each job has new, dedicated resources.&nbsp;&nbsp;</p>



<h4 class="wp-block-heading">ODP Cluster Security&nbsp;</h4>



<p>ODP will take care of the security and privacy of your cluster as well. Firstly, all communications between the Spark nodes are encrypted. Secondly, None of your job’s nodes are accessible from the outside. ODP only allows limited ports to be open for your cluster, so that you are still able to load or push your data.&nbsp;</p>



<h4 class="wp-block-heading">ODP Cluster Spark Version&nbsp;</h4>



<p>When it comes to using multiple Spark versions on the same cluster, ODP offers a solution. As every job possesses its own dedicated resources, each job can use any version currently supported by the service, independently from any other job running at&nbsp; the same time. When submitting an Apache Spark job through ODP, you will first select the version of Apache Spark you would like to use. When the Apache Spark community releases a new version, it will soon become available in ODP and you can then submit another job with the new Spark version as well. This means you don’t need to keep updating the Spark version of your whole cluster anymore.&nbsp;</p>



<h4 class="wp-block-heading">ODP Cluster Efficiency</h4>



<p>Each time you submit a job, you’ll have to define exactly how many resources and workers you would like to use for that job. As said earlier, each job has its own dedicated resources so you will be able to have small jobs running alongside much bigger ones. This flexibility, means that you will never have to worry about having an idle cluster. You pay for the resources you use, when you use them.</p>



<div class="wp-block-image"><figure class="aligncenter size-large is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2019/05/CD83F4F3-4858-414A-A50E-362477780385-1024x662.png" alt="OVHcloud Data Processing" class="wp-image-15537" width="768" height="497" srcset="https://blog.ovhcloud.com/wp-content/uploads/2019/05/CD83F4F3-4858-414A-A50E-362477780385-1024x662.png 1024w, https://blog.ovhcloud.com/wp-content/uploads/2019/05/CD83F4F3-4858-414A-A50E-362477780385-300x194.png 300w, https://blog.ovhcloud.com/wp-content/uploads/2019/05/CD83F4F3-4858-414A-A50E-362477780385-768x496.png 768w, https://blog.ovhcloud.com/wp-content/uploads/2019/05/CD83F4F3-4858-414A-A50E-362477780385-1200x775.png 1200w, https://blog.ovhcloud.com/wp-content/uploads/2019/05/CD83F4F3-4858-414A-A50E-362477780385.png 2048w" sizes="auto, (max-width: 768px) 100vw, 768px" /></figure></div>



<h3 class="wp-block-heading">How to start?&nbsp;</h3>



<p>If you&#8217;re interested in trying ODP, you can check out: <a href="https://www.ovhcloud.com/en/public-cloud/data-processing/" data-wpel-link="external" target="_blank" rel="nofollow external noopener noreferrer">https://www.ovhcloud.com/en/public-cloud/data-processing/</a> or you can easily create an account at <a href="http://www.ovhcloud.com" data-wpel-link="external" target="_blank" rel="nofollow external noopener noreferrer">www.ovhcloud.com</a> and select “data processing” in the public cloud section. It is also possible to ask questions directly from the product team in the ODP public gitter channel <a href="https://gitter.im/ovh/data-processing" data-wpel-link="external" target="_blank" rel="nofollow external noopener noreferrer">https://gitter.im/ovh/data-processing</a>.</p>



<h3 class="wp-block-heading">Conclusion&nbsp;</h3>



<p>With ODP, the challenges of running an Apache Spark cluster are removed, or alleviated (we still can’t do much about users’ expectations!) You don’t have to worry about the lack of resources necessary to process your data, or the need to create, install and manage your own cluster. </p>



<p>Focus on your processing algorithm and ODP will do the rest.</p>
<img loading="lazy" decoding="async" src="//blog.ovhcloud.com/wp-content/plugins/matomo/app/matomo.php?idsite=1&amp;rec=1&amp;url=https%3A%2F%2Fblog.ovhcloud.com%2Fwhy-are-you-still-managing-your-data-processing-clusters%2F&amp;action_name=Why%20are%20you%20still%20managing%20your%20data%20processing%20clusters%3F&amp;urlref=https%3A%2F%2Fblog.ovhcloud.com%2Ffeed%2F" style="border:0;width:0;height:0" width="0" height="0" alt="" />]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Improving the quality of data with Apache Spark</title>
		<link>https://blog.ovhcloud.com/improving-the-quality-of-data-with-apache-spark/</link>
		
		<dc:creator><![CDATA[Hubert Stefani]]></dc:creator>
		<pubDate>Tue, 15 Sep 2020 15:34:26 +0000</pubDate>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[Apache Spark]]></category>
		<category><![CDATA[Data Processing]]></category>
		<category><![CDATA[OVHcloud Data Processing]]></category>
		<guid isPermaLink="false">https://www.ovh.com/blog/?p=18676</guid>

					<description><![CDATA[Today we are proposing you a guest post by Hubert Stefani, Chief Innovation Officer and Cofounder of Novagen Conseil As data consultant experts and heavy Apache Spark users, we felt honoured to become early adopters of OVHcloudData Processing. As a first use case to test this offering, we chose our quality assessment process. As a [&#8230;]<img src="//blog.ovhcloud.com/wp-content/plugins/matomo/app/matomo.php?idsite=1&amp;rec=1&amp;url=https%3A%2F%2Fblog.ovhcloud.com%2Fimproving-the-quality-of-data-with-apache-spark%2F&amp;action_name=Improving%20the%20quality%20of%20data%20with%20Apache%20Spark&amp;urlref=https%3A%2F%2Fblog.ovhcloud.com%2Ffeed%2F" style="border:0;width:0;height:0" width="0" height="0" alt="" />]]></description>
										<content:encoded><![CDATA[
<div class="wp-block-group"><div class="wp-block-group__inner-container is-layout-flow wp-block-group-is-layout-flow">
<p><em>Today we are proposing you a guest post by</em> Hubert Stefani, Chief Innovation Officer and Cofounder of <a href="http://www.novagen.tech/" data-wpel-link="external" target="_blank" rel="nofollow external noopener noreferrer">Novagen Conseil</a></p>
</div></div>



<figure class="wp-block-image size-large is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2020/09/IMG_0269-1024x537.png" alt="Improving the quality of data with Apache Spark" class="wp-image-19307" width="768" height="403" srcset="https://blog.ovhcloud.com/wp-content/uploads/2020/09/IMG_0269-1024x537.png 1024w, https://blog.ovhcloud.com/wp-content/uploads/2020/09/IMG_0269-300x157.png 300w, https://blog.ovhcloud.com/wp-content/uploads/2020/09/IMG_0269-768x403.png 768w, https://blog.ovhcloud.com/wp-content/uploads/2020/09/IMG_0269.png 1200w" sizes="auto, (max-width: 768px) 100vw, 768px" /></figure>



<p>As data consultant experts and heavy Apache Spark users, we felt honoured to become early adopters of <a href="https://www.ovhcloud.com/en-ie/public-cloud/data-processing/" data-wpel-link="external" target="_blank" rel="nofollow external noopener noreferrer">OVHcloudData Processing</a>. As a first use case to test this offering, we chose our quality assessment process.</p>



<p>As a data consultancy company based in Paris, we build complete and innovative data strategies for our large corporate and public customers: the top fortune banks, public authorities, retailers, fashion industry, transportation leaders etc. We offer them huge scale BI, data lake creation and management, business innovation with data science. Within our Data Lab, we select the best-in-class technology and create what we call ‘boosters’ ie ready to-deploy or customized data assets.</p>



<p>When it comes to selecting a new technology solution, we have the following check list:</p>



<ul class="wp-block-list"><li><strong>Innovation and evolutivity</strong>: depth of functionalities, additional value and usability</li><li><strong>Performance and cost-effectiveness</strong>: intrinsic performances, but also technical architectures that adapt to customer needs</li><li><strong>Open standards and governance</strong>: to support our customers’ cloud or multi-cloud strategies, we choose to rely on open standards to deploy on different targets and preserve reversibility.</li></ul>


<div class="lazyblock-youtube-gdpr-compliant-Z2pdIhN wp-block-lazyblock-youtube-gdpr-compliant"><script type="module">
  import 'https://blog.ovhcloud.com/wp-content/assets/ovhcloud-gdrp-compliant-embedding-widgets/src/ovhcloud-gdrp-compliant-spreaker.js';
</script>
      
      <ovhcloud-gdrp-compliant-spreaker
          spreaker=""
          debug></ovhcloud-gdrp-compliant-spreaker> 

</div>


<h3 class="wp-block-heading"> Apache Spark, our Swiss Army knife</h3>



<p>About a month ago OVHcloud’s Data and AI Product Manager, Bastien Verdebout approached us to test its new product OVHcloud Data Processing, built on top of Apache Spark. The answer was of course yes!</p>



<p>One of the reasons we felt so eager to discover this data processing as a service solution was that we have an extensive usage of Apache Spark; it’s our our Swiss Army knife to process data.</p>



<ul class="wp-block-list"><li>It works on extremely high scale of data,</li><li>It meets the needs of data engineering and data science,</li><li>It allows the processing of data at rest and data streaming</li><li>It’s the de facto standard for data workloads on-premises and in the Cloud,</li><li>It offers built-in APIs for Python, Scala, Java and R.</li></ul>



<p>We have progressively developped software assets on top of Apache Spark to address recurring challenges such as:</p>



<ul class="wp-block-list"><li>ETL processing in data lake environnements,</li><li>Quality KPIs on top of data lake sources,</li><li>Machine Learning algorithm for Natural Language Processing, Time Series predictions&#8230;</li></ul>



<h3 class="wp-block-heading">The ideal use case: data quality assessment</h3>



<p>We have considered the following charateristics of<strong> OVHCloud Data processing</strong>:</p>



<ul class="wp-block-list"><li>Processing engine built on top of<strong> Apache Spark 2.4.3</strong></li><li>Jobs start after <strong>a few seconds</strong> (vs minutes to launch a cluster)</li><li>Ability to<strong> adjust power dedicated to different Spark jobs</strong>: start with low power (1 driver and 1 executor with 4 cores and 8Gb of memory) to high scale processing (potential hundreds of cores and Gb of memories)</li><li>A full <strong>Compute/Storage separation</strong> aligned with <strong>standard of cloud architectures</strong>, including S3 APIs to access data stored in Object Storage layer. &nbsp; </li><li>Jobs execution and monitoring through <strong>Command Line Interface</strong> and <strong>API</strong>&nbsp;</li></ul>



<p>These characteristics led us to choose our quality assessment process as an ideal use case which requires both interactivity and adjustable compute resources to deliver quality KPIs through Spark processes.</p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="752" src="https://www.ovh.com/blog/wp-content/uploads/2020/09/IMG_0268-1024x752.png" alt="Why we need Spark as a Service" class="wp-image-19299" srcset="https://blog.ovhcloud.com/wp-content/uploads/2020/09/IMG_0268-1024x752.png 1024w, https://blog.ovhcloud.com/wp-content/uploads/2020/09/IMG_0268-300x220.png 300w, https://blog.ovhcloud.com/wp-content/uploads/2020/09/IMG_0268-768x564.png 768w, https://blog.ovhcloud.com/wp-content/uploads/2020/09/IMG_0268.png 1497w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<h3 class="wp-block-heading"> OVHCloud Data Processing at work</h3>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="960" height="540" src="https://www.ovh.com/blog/wp-content/uploads/2020/07/image-blog-post-Novagen-2.png" alt="" class="wp-image-18981" srcset="https://blog.ovhcloud.com/wp-content/uploads/2020/07/image-blog-post-Novagen-2.png 960w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/image-blog-post-Novagen-2-300x169.png 300w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/image-blog-post-Novagen-2-768x432.png 768w" sizes="auto, (max-width: 960px) 100vw, 960px" /></figure>



<p>The corresponding command generated by our software is:</p>



<pre class="wp-block-code"><code class="">./ovh-spark-submit --projectid ec7d2cb6da084055a0501b2d8d8d62a1 \
  --class tech.novagen.spark.Launcher --driver-cores 4 --driver-memory 8G \
  --executor-cores 4 --executor-memory 8G --num-executors 5 \ 
  swift://sparkjars/QualitySparkExecutor-1.0-spark.jar --apiServer=5.1.1.2:80</code></pre>



<p>We have a command which is quite similar to a usual spark-submit, except for the jar path, which requires the binary to be in an Object Storage bucket that we access with swift url specification. (NB: this command could have been created with a call to the OVHCloud Data Processing API).</p>



<p>Starting from this point, we can now fine tune our process portfolio and play with the allocation of different power with little limitation (except for the quotas assigned to any public cloud project).</p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="960" height="540" src="https://www.ovh.com/blog/wp-content/uploads/2020/07/image-blog-post-Novagen-3.png" alt="" class="wp-image-18982" srcset="https://blog.ovhcloud.com/wp-content/uploads/2020/07/image-blog-post-Novagen-3.png 960w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/image-blog-post-Novagen-3-300x169.png 300w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/image-blog-post-Novagen-3-768x432.png 768w" sizes="auto, (max-width: 960px) 100vw, 960px" /></figure>



<h2 class="wp-block-heading"> Real-time display of job logs</h2>



<p>In the end, for tuning and post-mortem job analysis, we can take advantage of the saved log files. It is noteworthy that OVHcloud Data Processing offers a real time display of job logs which is very convenient and provides a complementary supervision through Grafana dashboards.</p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="960" height="540" src="https://www.ovh.com/blog/wp-content/uploads/2020/07/image-blog-post-Novagen-4.png" alt="" class="wp-image-18983" srcset="https://blog.ovhcloud.com/wp-content/uploads/2020/07/image-blog-post-Novagen-4.png 960w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/image-blog-post-Novagen-4-300x169.png 300w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/image-blog-post-Novagen-4-768x432.png 768w" sizes="auto, (max-width: 960px) 100vw, 960px" /></figure>



<p>This is a first yet significant test of <a href="https://www.ovhcloud.com/en-ie/public-cloud/data-processing/" data-wpel-link="external" target="_blank" rel="nofollow external noopener noreferrer">OVHcloud Data Processing</a>; so far, it proved an excellent match with the Novagen quality process use case and allowed us to validate several crucial points when it comes to testing a new data solution:</p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="960" height="540" src="https://www.ovh.com/blog/wp-content/uploads/2020/07/image-blog-post-Novagen-5.png" alt="" class="wp-image-18984" srcset="https://blog.ovhcloud.com/wp-content/uploads/2020/07/image-blog-post-Novagen-5.png 960w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/image-blog-post-Novagen-5-300x169.png 300w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/image-blog-post-Novagen-5-768x432.png 768w" sizes="auto, (max-width: 960px) 100vw, 960px" /></figure>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow"><p>This is the beginning of this product, and we will have a close look at the upcoming functionalities. The OVHCloud team unveiled part of its roadmap, and it looks really promising!</p><cite>Hubert Stefani, Chief Innovation Officer of Novagen Conseil</cite></blockquote>



<p></p>
<img loading="lazy" decoding="async" src="//blog.ovhcloud.com/wp-content/plugins/matomo/app/matomo.php?idsite=1&amp;rec=1&amp;url=https%3A%2F%2Fblog.ovhcloud.com%2Fimproving-the-quality-of-data-with-apache-spark%2F&amp;action_name=Improving%20the%20quality%20of%20data%20with%20Apache%20Spark&amp;urlref=https%3A%2F%2Fblog.ovhcloud.com%2Ffeed%2F" style="border:0;width:0;height:0" width="0" height="0" alt="" />]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Do you need to process your data? Try the new OVHcloud Data Processing service!</title>
		<link>https://blog.ovhcloud.com/try-the-new-ovhcloud-data-processing-service/</link>
		
		<dc:creator><![CDATA[Mojtaba Imani]]></dc:creator>
		<pubDate>Wed, 22 Jul 2020 12:55:45 +0000</pubDate>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[Apache Spark]]></category>
		<category><![CDATA[Big Data]]></category>
		<category><![CDATA[Cluster Computing]]></category>
		<category><![CDATA[Data Platform]]></category>
		<category><![CDATA[Data Processing]]></category>
		<category><![CDATA[Distributed Computing]]></category>
		<category><![CDATA[OVHcloud]]></category>
		<guid isPermaLink="false">https://www.ovh.com/blog/?p=18748</guid>

					<description><![CDATA[One of the data services of OVHcloud is called OVHcloud Data Processing (ODP). It is a service that allows you to submit a processing job without caring about the cluster behind it. You just have to specify the ressources you want to use for your job, and the service will abstract the cluster creation, and destroy it for you as soon as your job is finished. In other words, you don’t have to think about clusters any more. Decide how much resources you need to process your data in the most efficient way for you and let OVHcloud Data Processing do the rest.<img src="//blog.ovhcloud.com/wp-content/plugins/matomo/app/matomo.php?idsite=1&amp;rec=1&amp;url=https%3A%2F%2Fblog.ovhcloud.com%2Ftry-the-new-ovhcloud-data-processing-service%2F&amp;action_name=Do%20you%20need%20to%20process%20your%20data%3F%20Try%20the%20new%20OVHcloud%20Data%20Processing%20service%21&amp;urlref=https%3A%2F%2Fblog.ovhcloud.com%2Ffeed%2F" style="border:0;width:0;height:0" width="0" height="0" alt="" />]]></description>
										<content:encoded><![CDATA[
<p>Today, we are generating more data than ever. 90 percent of global data has been generated in the last 2 years. By 2025 the amount of data in the world is estimated to reach 175 Zettabyte. In total, people write 500 million tweets per day, and autonomous cars generate 20TB of data every hour. By the year 2025 more than 75 billion IoT devices will be connected to the web, which will all generate data. Nowadays, devices and services that generate data are everywhere. .</p>



<div class="wp-block-image"><figure class="aligncenter size-large is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2020/07/DAF95617-1C6B-49FC-A2CD-8346B1E6B9AC-1024x539.jpeg" alt="Do you need to process your data? Try the new OVHcloud Data Processing cloud service!" class="wp-image-18852" width="512" height="270" srcset="https://blog.ovhcloud.com/wp-content/uploads/2020/07/DAF95617-1C6B-49FC-A2CD-8346B1E6B9AC-1024x539.jpeg 1024w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/DAF95617-1C6B-49FC-A2CD-8346B1E6B9AC-300x158.jpeg 300w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/DAF95617-1C6B-49FC-A2CD-8346B1E6B9AC-768x404.jpeg 768w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/DAF95617-1C6B-49FC-A2CD-8346B1E6B9AC.jpeg 1200w" sizes="auto, (max-width: 512px) 100vw, 512px" /></figure></div>



<p>There is also the notion of data exhaust which is the by-product of people&#8217;s online activities. It’s the data that’s generated as a result of someone visiting a website, buying a product or searching for something using a search engine. You may have heard this data described as Metadata.</p>



<p>We will start drowning in a flood of data unless we learn how to swim &#8211; how to benefit from vast amounts of data. To do this we need to be able to process the data for the sake of better decision-making, preventing fraud and danger,&nbsp; inventing better products or even predicting the future. The possibilities are endless.</p>



<p>But how can we process this huge amount of data? For sure, it’s not possible to do it the old fashioned way. We need to upgrade our methods and equipment.</p>



<p>Big data are data sets that have a volume, velocity and variety too large to be processed by a local computer. So what are the requirements for processing “big data”? </p>



<div class="wp-block-image"><figure class="aligncenter size-large is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2020/07/7127B735-8372-4063-97DF-84BF3172E49A.jpeg" alt="Data, data everywhere!" class="wp-image-18850" width="436" height="368" srcset="https://blog.ovhcloud.com/wp-content/uploads/2020/07/7127B735-8372-4063-97DF-84BF3172E49A.jpeg 872w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/7127B735-8372-4063-97DF-84BF3172E49A-300x253.jpeg 300w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/7127B735-8372-4063-97DF-84BF3172E49A-768x648.jpeg 768w" sizes="auto, (max-width: 436px) 100vw, 436px" /></figure></div>



<h3 class="wp-block-heading">1- Process data in parallel</h3>



<p>Data is everywhere and available in huge quantities. First off, lets apply the old rule: &#8216;Divide and Conquer&#8217;.&nbsp;</p>



<p>Dividing the data means that we will need data and processing tasks to be distributed across several computers. These will need to be set in a cluster, to perform these different tasks in parallel and gain reasonable performance and speed boosts.</p>



<p>Lets assume that you needed to find out whats trending on twitter. You would have to process around 500 million tweets with one computer in one hour. Not so easy now, is it? And how would you benefit if it took a month to process? What is the value in finding the trend of the day, one month later?</p>



<p>Parallelization is more than a “nice to have” feature. It’s a requirement!</p>



<h3 class="wp-block-heading">2- Process data in the cloud</h3>



<p>The second step is to create and manage these clusters in an efficient way.&nbsp;</p>



<p>You have several choices here, like creating clusters with your own servers and managing them by yourself. But that’s time consuming and costs quite a lot as well. It also lacks some features you may wish to use, like flexibility. For these reasons, the cloud appears to be a better and better solution every day for a lot of companies.</p>



<p>The elasticity that cloud solutions provide, helps companies to be flexible and adapt infrastructure to their needs. With data processing, for example, we will need to be able to scale up and down our computing cluster easily to adapt the computing power to the volume of data we want to process according to our constraints (time, costs, etc.).&nbsp;</p>



<p>And then, even if you decide to use a cloud provider, you will have several solutions to choose from, each with their own drawbacks. One of these solutions, is to create a computing cluster over the dedicated servers or public cloud instances and send different processing jobs to the cluster. The main drawback, in this solution, would be that if no processing is done, you’re still paying for the reserved but unused processing resources.&nbsp;</p>



<p>A more efficient way would therefore be to create a dedicated cluster for each processing job, with the right resources for that job, and then delete the cluster after. Each new job would have it’s own cluster, sized as needed, spawned on demand. But this solution would only be feasible if the creation of a computing cluster took but a few seconds and not minutes or hours.</p>



<h4 class="wp-block-heading">Data locality</h4>



<p>When creating a processing cluster, it is also possible to consider data locality. Here, cloud providers usually offer several regions spread across data centers situated in different countries. It has two main benefits: </p>



<p>The first one is not directly linked to data locality but more of a legal point. Depending on where your customers are, and where your data is, you may need to comply with local data privacy laws and regulations. You may need to keep your data in a specific region or country and not be able to process it outside. So, to create a cluster of computers in that region, it is easier to process data while complying with local privacy policies.</p>



<p>The second benefit is, of course, the potential to create your processing clusters in close physical proximity to your data. According to estimations, by the year 2025, almost 50 percent of data in the world will be stored in the cloud. On-premises data storing is in decline. </p>



<p class="has-text-align-center"><img loading="lazy" decoding="async" src="https://lh5.googleusercontent.com/xLg8hqGdVDbuf4AbWODV95LN9DInueit-Lqv6xxjz61lGKJ4ukK7noRAcpiLKrm5To8-ztAstlFqfcq1qZqMerZwg1wYjMcxtQmXvqpwZzVemn9vz1lPowxb8h3CA7O0ug0C_GQa" width="624" height="360"></p>



<p>Therefore, using cloud providers that have several regions gives companies the benefit of having processing clusters near their datas physical location &#8211; this greatly reduces the time (and costs!) it takes to fetch data.</p>



<p>While processing your data in the cloud may not be a requirement per se, its certainly more beneficial than doing it yourself.</p>



<h3 class="wp-block-heading">3- Process data with the most appropriate distributed computing technology</h3>



<p>The third and final step, is to decide how you are going to process your data, meaning with what tools. Again, you could do it by yourself, by implementing a distributed processing engine in a language of your choice. But where&#8217;s the fun in that? (okay, for some of us it might actually be quite fun!)</p>



<p>But it would be astronomically complex. You would need to write code to divide the data into several parts and send each part to a computer in your cluster. Each computer would then process its part of data, and you would need to find a way to retrieve the results of each part and to re-aggregate everything into a coherent result. In short, it would be a lot of work, with a lot of debugging.&nbsp;</p>



<h4 class="wp-block-heading">Apache Spark</h4>



<p>But there are technologies that have been developed specifically for this purpose. They distribute the data and processing tasks automatically and retrieve the results for you. Currently, the most popular distributed computing technology, especially in relation to Data Science subjects, is Apache Spark.</p>



<p>Apache Spark is an open-source, distributed, cluster-computing framework. It is much faster than the previous one, Hadoop MapReduce, thanks to features like in-memory processing and lazy evaluation.</p>



<p>Apache Spark is the leading platform for large-scale SQL, batch processing, stream processing and machine learning. For coding in Apache Spark, you have the option of using different programming languages (including Java, Scala, Python, R and SQL). It can run locally in a single machine or in a cluster of computers to distribute its tasks.</p>



<figure class="wp-block-image"><img decoding="async" src="https://lh5.googleusercontent.com/tHvP_t-ZFlfnCy00xVoj6hu9o_SPWa3b2Wl9O-CdRKXOw8ZF3fDqr9BnDppvAfgPu34wi_eRg4mRnBYbCcvdqK9SxHaV7vAh_stEbp4yuxIUpnujmf24gA3tADPdUNxjdUXSoyUs" alt=""/></figure>



<p>As you can see in the google trends data chart above, there are alternatives. But Apache Spark has definitely established itself as the leader in distributed computing tools.</p>



<h3 class="wp-block-heading">OVHcloud Data Processing (ODP)</h3>



<p>OVHcloud is the leader of European hosting and cloud providers with a wide range of cloud services like public and private cloud, managed Kubernetes and cloud storage. But besides all the hosting and cloud services, OVHcloud also provides a range of big data analytics and artificial intelligence services as well as platforms. </p>



<p>One of the data services offered by OVHcloud is OVHcloud Data Processing (ODP). It is a service that allows you to submit a processing job without worrying about the cluster behind it. You just have to specify the resources you need for the job and the service will abstract the cluster creation, and destroy it for you as soon as your job finishes. In other words, you don’t have to think about clusters any more. Decide on how many resources you will need to process your data in an efficient way, and then let OVHcloud Data Processing do the rest.&nbsp;</p>



<h4 class="wp-block-heading">On-demand, job-specific Spark clusters</h4>



<p>The service will deploy a temporary, job-specific Apache Spark Cluster, then configure and secure it automatically. You don’t need to have any prior knowledge or skills related to the cloud, networking, cluster management systems, security, etc. You only have to focus on your processing algorithm and Apache Spark code.</p>



<p>This service will download your Apache Spark code from one of your Object Storage containers, and ask you how much RAM and CPU cores you would like your job to use. You will also have to specify the region you want the processing to take place in. Last but not least, you will then have to choose the Apache Spark version you want to use to run your code. The service will then launch your job within a few seconds, according to the specified parameters until your job’s completion. Nothing else to do on your part. No cluster creation, no cluster destruction. Just focus on your code.</p>



<p>Your local computer resources no longer limit the amount of data you can process. You can run any number of processing jobs in parallel in any region and any version of Spark. Its also very fast and very easy.</p>



<div class="wp-block-image"><figure class="aligncenter size-large is-resized"><img loading="lazy" decoding="async" src="https://www.ovh.com/blog/wp-content/uploads/2020/07/EF2587E4-9E30-4C3B-A6E1-0EAF3FF8F8F6-1024x813.jpeg" alt="OVHcloud Data Processing" class="wp-image-18856" width="512" height="407" srcset="https://blog.ovhcloud.com/wp-content/uploads/2020/07/EF2587E4-9E30-4C3B-A6E1-0EAF3FF8F8F6-1024x813.jpeg 1024w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/EF2587E4-9E30-4C3B-A6E1-0EAF3FF8F8F6-300x238.jpeg 300w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/EF2587E4-9E30-4C3B-A6E1-0EAF3FF8F8F6-768x610.jpeg 768w, https://blog.ovhcloud.com/wp-content/uploads/2020/07/EF2587E4-9E30-4C3B-A6E1-0EAF3FF8F8F6.jpeg 1227w" sizes="auto, (max-width: 512px) 100vw, 512px" /></figure></div>



<h3 class="wp-block-heading">How does it work ?</h3>



<p>On your side You just need to:&nbsp;</p>



<ol class="wp-block-list"><li>Create a container in OVHcloud Object Storage and upload the Apache Spark code and any other required files to this container. Be careful not to put your data in the same container as well, as the whole container will be downloaded by the service.</li><li>You then have to define the processing engine (like Apache Spark) and its version, as well as the geographical region and the amount of resources (CPU cores, RAM and number of worker nodes) you need. There are three different ways to execute this (OVHcloud Control Panel, API or ODP CLI)</li></ol>



<p>These are different steps that happen when you run a processing job in OVHcloud Data Processing (ODP) platform: <strong>&nbsp;</strong></p>



<div class="wp-block-group"><div class="wp-block-group__inner-container is-layout-flow wp-block-group-is-layout-flow">
<ol class="wp-block-list"><li>ODP will take over and handle the deployment and execution of your job according to the specifications that you defined.</li><li>Before starting your job, ODP will download all files that you uploaded in the specified container.</li><li>Next, ODP will run your job in a dedicated environment, created specifically for your job. Apart from a limitation on the available ports (list <a href="https://docs.ovh.com/gb/en/data-processing/capabilities/" data-wpel-link="exclude">available here</a>), your job can then connect to any data source (databases, object storage, etc) to read or write data (as long as they are reachable through the Internet)</li><li>When the job is complete, ODP stores the execution output logs to your Object Storage and then deletes the whole cluster immediately.</li><li>You will be charged for the amount of resources you specified and only for the duration of your job computation, on a per-minutes basis.</li></ol>
</div></div>



<h3 class="wp-block-heading">Different ways to submit a job?</h3>



<p>There are three different ways that you can submit a processing job to ODP, depending on your requirements. These three ways are OVHcloud Manager, OVHcloud API and CLI (Command Line Interface).&nbsp;</p>



<h4 class="wp-block-heading">1. OVHcloud Manager</h4>



<p>To submit a job with OVHcloud Manager you need to go to OVHcloud.com and login with your OVHcloud account (or create one if neccessary). Then go to the “Public Cloud” page and select the “Data Processing” link on the left panel and submit a job by clicking on “Start a new job”.</p>



<p>Before submitting a job you need to create a container in OVHcloud Object Storage by clicking on “Object Storage” link on the left panel and upload your Apache Spark code and any other required files.&nbsp;</p>



<h4 class="wp-block-heading">2. OVHcloud API&nbsp;</h4>



<p>You can submit a job to ODP by using OVHcloud API. For more information, you can see the OVHcloud API web page <a href="https://api.ovh.com/" data-wpel-link="exclude">https://api.ovh.com/</a>. You can create job submit automation by using ODP API.&nbsp;</p>



<h4 class="wp-block-heading">3. ODP CLI (Command Line Interface)&nbsp;</h4>



<p>ODP has an open source Command Line Interface that you can find in OVH public GitHub at <a href="https://github.com/ovh/data-processing-spark-submit" data-wpel-link="external" target="_blank" rel="nofollow external noopener noreferrer">https://github.com/ovh/data-processing-spark-submit</a>). By using CLI, you can upload your files and codes and create your Apache Spark cluster together with just one command.&nbsp;</p>



<h3 class="wp-block-heading">Some ODP benefits</h3>



<p>You can either always run your processing tasks in your local computer, or you can create an Apache Spark cluster in your local premises with any cloud provider. This means you can manage that cluster yourself or using similar services from other competitors. But ODP has several benefits, it is good to have them in your mind when deciding on a solution:</p>



<ul class="wp-block-list"><li><strong>No cluster management or configuration</strong> skills or experience is needed.</li><li><strong>Not limited by resources</strong> and easy and fast. (The only limit is your cloud account quota)&nbsp;</li><li><strong>Pay as you go </strong>model with easy pricing and no hidden cost. (per-minutes billing)</li><li><strong>Per job resource definition</strong> (no more resources lost compared to a mutualised cluster)&nbsp;</li><li><strong>Ease of managing Apache Spark version </strong>(You select the version for each job and you can even have different jobs with different versions of the Apache Spark at the same time)&nbsp;</li><li><strong>Region</strong> <strong>selection</strong> (You can select different regions based on your data locality or data privacy policy)</li><li><strong>Start a Data Processing job</strong> <strong>in just a few seconds</strong></li><li><strong>Real-time logs </strong>(when your job is running, you will receive <strong>real-time logs</strong> in your Customer Panel)</li><li><strong>Full output log will be available just after finishing the job </strong>(some competitors take minutes to deliver logs to you)</li><li><strong>Job submit automation</strong> (by using ODP API or CLI)</li><li><strong>Data Privacy </strong>(OVHcloud is European company and all customers are strictly protected by European GDPR)</li></ul>



<h3 class="wp-block-heading">Conclusion</h3>



<p>With the advance of new technologies and devices, we are flooded with data. More and more, it is essential for businesses and for academic research to process data sets and understand where the value is. By providing the OVHcloud Data Processing (ODP) service, our goal is to provide you with one of the easiest and most efficient platforms to process your data. Just focus on your processing algorithm and ODP will handle the rest for you.</p>
<img loading="lazy" decoding="async" src="//blog.ovhcloud.com/wp-content/plugins/matomo/app/matomo.php?idsite=1&amp;rec=1&amp;url=https%3A%2F%2Fblog.ovhcloud.com%2Ftry-the-new-ovhcloud-data-processing-service%2F&amp;action_name=Do%20you%20need%20to%20process%20your%20data%3F%20Try%20the%20new%20OVHcloud%20Data%20Processing%20service%21&amp;urlref=https%3A%2F%2Fblog.ovhcloud.com%2Ffeed%2F" style="border:0;width:0;height:0" width="0" height="0" alt="" />]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
