{"id":303,"date":"2017-07-11T04:57:49","date_gmt":"2017-07-11T04:57:49","guid":{"rendered":"http:\/\/eurekadata.net\/?page_id=303"},"modified":"2017-07-11T15:21:27","modified_gmt":"2017-07-11T15:21:27","slug":"size","status":"publish","type":"page","link":"https:\/\/eurekadata.net\/index.php\/introduction\/size\/","title":{"rendered":"The Illusion Resource Control"},"content":{"rendered":"<p class=\"p1\">The very notion of <i>big data<\/i> defines <em>data<\/em>, whatever that is by its being <em>big<\/em> whatever that is. \u00a0The abstraction of fields in records, or by the combined size or perhaps utilizing statistical distribution of the values in those fields only exacerbates the ambiguity. \u00a0Qualifying order of magnitude numbers is difficult enough as they go above the countable range, but with format or structural changes or codec the same data can vary enormously in size.<\/p>\n<p class=\"p1\">In practice big data practitioners typically resort to raw amount of total space utilized. \u00a0If this sounds at all reasonable please consider the following. \u00a0Given that I could take three hundred gallons to paint my house, one could derive drastically different conclusions. \u00a0On the one hand, this figure could indicate that I am the owner of the possibly the largest and most expensive property in my city. \u00a0On the other hand, it could indicate a terrible painter.<\/p>\n<p class=\"p1\">I have claimed and we can reasonably deduce that the meta data in any big data system is a good multiple of the resources of original data. \u00a0Of the big data technology stack one feature regularly presents itself is opportunity to tune performance or coverage against resource allocation. \u00a0This sounds reasonable on the surface as in practice decisions result in significant quantities of resources, but what if the base line cost is not very reasonable.<\/p>\n<p class=\"p1\">I examine some actual numbers in the <a href=\"http:\/\/eurekadata.net\/index.php\/performance\/\">performance section<\/a> but insofar the a la cart approach used in most big data installations are inefficient in contrast to a wholistic approach of Eureka then the opportunity to pare down expensive options represents the illusion of economy.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>The very notion of big data defines data, whatever that is by its being big whatever that is. \u00a0The abstraction of fields in records, or by the combined size or perhaps utilizing statistical distribution of the values in those fields only exacerbates the ambiguity. \u00a0Qualifying order of magnitude numbers is difficult enough as they go &hellip; <a href=\"https:\/\/eurekadata.net\/index.php\/introduction\/size\/\" class=\"more-link\">Continue reading <span class=\"screen-reader-text\">The Illusion Resource Control<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"parent":238,"menu_order":4,"comment_status":"closed","ping_status":"closed","template":"","meta":[],"_links":{"self":[{"href":"https:\/\/eurekadata.net\/index.php\/wp-json\/wp\/v2\/pages\/303"}],"collection":[{"href":"https:\/\/eurekadata.net\/index.php\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/eurekadata.net\/index.php\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/eurekadata.net\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/eurekadata.net\/index.php\/wp-json\/wp\/v2\/comments?post=303"}],"version-history":[{"count":8,"href":"https:\/\/eurekadata.net\/index.php\/wp-json\/wp\/v2\/pages\/303\/revisions"}],"predecessor-version":[{"id":343,"href":"https:\/\/eurekadata.net\/index.php\/wp-json\/wp\/v2\/pages\/303\/revisions\/343"}],"up":[{"embeddable":true,"href":"https:\/\/eurekadata.net\/index.php\/wp-json\/wp\/v2\/pages\/238"}],"wp:attachment":[{"href":"https:\/\/eurekadata.net\/index.php\/wp-json\/wp\/v2\/media?parent=303"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}