How TripleBlind’s Data Privacy Solution Compares to Differential Privacy

Differential privacy is a privacy guarantee that allows for publicly sharing information about a dataset by describing the patterns of groups within the dataset while withholding information about individuals in the dataset.

When computing on data via differential privacy, it adds stochastic noise to each data element that masks the actual data element. Stochastic refers to a variable process where the outcome involves some randomness and has some uncertainty. This results in significant accuracy degradation, whereas TripleBlind’s one-way encryption algorithms don’t add any noise to the dataset that would impair results. 

Differential privacy is suitable for situations with a higher degree of tolerance for error. For example, Apple keyboard suggestions – Apple doesn’t need to know exactly what you’re typing, but needs to know in general what people are typing to offer reasonable suggestions. 

Apple itself sets a strict limit on the number of contributions from a user in order to preserve their privacy. The reason is that the slightly biased noise used in differential privacy tends to average out over a large number of contributions, making it theoretically possible to determine information about a user’s activity over a large number of observations from a single user. It’s important to note that Apple doesn’t associate any identifiers with information collected using differential privacy.

The majority of the other approaches to data collaboration we’ve covered only work for tabular or columnar data; including homomorphic encryption, synthetic data, secure enclaves, tokenization and differential privacy. These approaches face severe challenges when it comes to producing high-performance, accurate models on complicated datasets like x-ray image analysis. However, TripleBlind is the solution to this problem since these images are obfuscated and encrypted – complying with HIPAA regulations.

 

 

If you’re interested in knowing more about how TripleBlind can safely and efficiently unlock privacy for you, please email contact@tripleblind.ai for a free demo. Don’t forget to follow TripleBlind on Twitter and LinkedIn for our latest updates. 

This is the final blog of our Competitor Blog Series where we compared TripleBlind’s technology to other approaches of data collaboration. If you missed the other blogs, you can check them out below!

Read other blogs in this series:

Business Agreements
Homomorphic Encryption
Synthetic Data
Blockchain
Tokenization, Masking and Hashing
Federated Learning

How TripleBlind’s Data Privacy Solution Compares to Tokenization, Masking and Hashing

Tokenization is the process of turning a piece of data, such as an account number, into a random string of characters called a token that has no meaningful value if breached. Tokens serve as a reference to the original data, but cannot be used to guess those values. 

Its use is gaining popularity, especially in the financial services industry. However, there are several limitations to this approach to data sharing compared to TripleBlind. 

When you tokenize a particular data element, you’ve lost the ability to compute on that data element. Let’s say you’re tokenizing a social security number; it will make aggregation and dataset joining tasks much more difficult because the same social security number can be stored as different data types in different datasets resulting in different token values.

However, with TripleBlind, your end result has higher accuracy with 100% data fidelity because all elements in the data are used for computation. Nothing is hidden, removed, or replaced. The data is used as-is while in complete compliance with the strictest regulations (such as GDPR, CCPA, and HIPAA). 

Let’s say you try a different but similar approach – masking or hashing. Masking has various approaches ranging from simple to complex. A simple method is to replace the real data with null or constant values. A slightly sophisticated approach would be to mask the data to retain the identity of the original data to preserve its analytical value. Masking always preserves the format, but there are risks of reidentification. 

A hash function is any function that can be used to map data of arbitrary size to fixed-size values. The values returned by a hash function are called hash values, hash codes, digests, or simply hashes. The values are usually used to index a fixed-size table called a hash table.

When masking or hashing medical data for an element like male or female, it isn’t that helpful because every instance of “male” will have the same value, and every instance of “female” will mask/hash to the same value. Therefore, you must remove the 18 HIPAA identifiers from the datasets entirely before its use. 

TripleBlind’s innovative solution allows all those HIPAA identifiers to remain in the dataset with a 0% chance of the data being reidentified at any point. These identifiers include important information for medical insights, such as biometric identifiers or facial images.

 

HIPAA Identifiers

1. Name 10. Account Number
2. Address 11. Certificate or License Number
3. Significant Dates 12. Vehicle Identifiers
4. Phone Numbers 13. Device Identifiers
5. Fax Numbers 14. Web URL
6. Email Address 15. IP Address
7. Social Security Number 16. Finger or Voice Print
8. Medical Record Number 17. Photographic Images
9. Health Plan Beneficiary Number 18. Other Characteristics that Could Uniquely Identify an Individual

 

Tokenization only works for tabular and columnar data, so most organizations end up combining different approaches like masking and tokenization to get the maximum value out of their data, but it doesn’t have to be this way. Our solution is a one-fits-all type.

To find out how TripleBlind works for your business, schedule a call or reach out for a free demo at contact@tripleblind.ai.

To learn more about how TripleBlind compares to other competitors and methods of data collaborations, follow us on LinkedIn and Twitter to be notified when we post the next installation in our Competitor Blog Series. Check out our previous blogs here!

How TripleBlind’s Data Privacy Solution Compares to Blockchain

​​Blockchain is a shared, immutable ledger that facilitates recording transactions and tracking assets in a business network. It’s most commonly associated with cryptocurrency, a record of transactions made in bitcoin or other cryptocurrencies that are maintained across several computers that are linked in a peer-to-peer network.

Blockchain has its advantages. It’s a great way to keep an audit trail of who might have done what to your data, but it’s not a good solution for data sharing in the long run. With blockchain, the data stored can still be accessible by certain individuals via a private key. 

With TripleBlind, all parties involved in data sharing will always know what is being done to their data. We provide audit trails of all operations, plus all parties must provide cryptographic consent to every operation done. There’s a fine-grained control of data and algorithm interactions where TripleBlind can manage individual attributes and record-level permissions on the data. This allows for accurate cryptographic auditability of every data and algorithm interaction without anyone ever seeing the raw data.

Sharing data through blockchain means it’s inherently public. It allows multiple tiers upstream and downstream to be transparent and highly visible – two words you don’t want associated with sensitive data.

Lastly, blockchain is not built for the future. It’s necessary to have an approach to data sharing that won’t come undone and leave businesses scrambling to use the next best solution. It’s costly, inefficient, and ineffective. 

TripleBlind is the future of data sharing and complies with the strictest of data privacy laws and regulations. It can be used around the globe, and our operations will automatically comply with local regulations such as GDPR since everything stays one-way encrypted during our process, and no one gets a copy of the raw data. 

To schedule a call or free demo to explore how TripleBlind can work for your business, please reach out to contact@tripleblind.ai. To keep up to date with our latest blogs, follow us on Twitter and LinkedIn!

Read other blogs in this series:

Business Agreements
Homomorphic Encryption
Synthetic Data
Tokenization, Masking and Hashing
Federated Learning
Differential Privacy

How TripleBlind Compares To Federated Learning

Federated Learning is a form of data collaboration in which companies use algorithms to monitor consumer activity and then use what is learned from that data to improve or create more efficient algorithms.

Google coined the term Federated Learning in 2016, and the company has since been at the forefront of AI training through this method. From a high level of abstraction, Federated Learning goes through the following steps:

  • A central server chooses an algorithm or statistical model to be trained. The server transmits the model to several data providers, often referred to as clients (consumers, devices, companies, etc.);
  • Each client trains the model on their data locally and shares updates with the server;
  • The server receives model updates from all clients and aggregates them into a single global model. The most common aggregation approach is averaging.

Federated Learning has the opportunity to be beneficial in both healthcare and financial markets, with the potential to create unbiased medians of large amounts of consumer information. In healthcare, trained models via Federated Learning can help with diagnosing rare diseases based on other patient data. In fintech, Federated Learning allows institutions to detect crime and risk factors within their collaboration network. 

Federated Learning only accesses the results and learnings based on the algorithms, which are then sent back to the server without sharing the actual data. It is meant to keep individual consumer data private. However, while Federated Learning allows for more privacy than has previously been possible with AI, there are downfalls when it comes to the model privacy and efficiency of collaboration through Federated Learning. 

Because Federated Learning requires each of the clients to train the model on their entire dataset locally, there is both a high computational load and high communication overhead.

When multiple parties collaborate through Federated Learning, the model through which the collaboration takes place is known to everyone involved, making it susceptible to several attacks that could lead to data leakage. Moreover, it also puts the actual model privacy at risk.

TripleBlind’s Blind Learning approach is superior to Federated Learning and is more computing and communication efficient. 

TripleBlind offers a more secure and precise way to share data. With TripleBlind’s groundbreaking solution, de-identified data is shared through models in which TripleBlind and all other parties involved are blind to the model and the original data.

 

 


Data sets are shared so that only relevant information to the collaboration can only be used for its intended purpose. By preventing reconstruction attacks, TripleBlind ensures there is no risk of the data ever being re-identified if there were to be a data breach.

We are comparing TripleBlind’s technology to other modes of data collaboration as part of our Competitor Blog Series. Stay up to date with TripleBlind on Twitter and LinkedIn to learn more. If you’re interested in knowing more about how collaborating using TripleBlind’s patented solution can safely and efficiently unlock privacy for you, please email contact@tripleblind.ai for a free demo.

Read other blogs in this series:

Business Agreements
Homomorphic Encryption
Synthetic Data
Blockchain
Tokenization, Masking and Hashing
Differential Privacy

How TripleBlind’s Data Privacy Solution Compares to Synthetic Data

Synthetic data is a form of collaboration in which businesses can share information with each other to analyze it without sharing real customer or patient information. An obvious downfall of collaborating by sharing synthetic data is that businesses are sharing generic data sets and not real data; however, synthetic data is acceptable when real data is unnecessary.

For example, synthetic data may be used by a credit card aggregator to determine macro trends from the data because not every bank collaborates with them and not every credit card provider will offer data. In those situations, synthetic data would be acceptable to glean industry macro-trends from the data.

However, if a company wanted to determine if a customer deserves a particular credit limit or understand how a small part of the population’s microtransactions yield a certain insight, they would need real data.

Another problem with sharing synthetic data is that outlying data is often omitted, making the data set inaccurate or can later be identified through spear-phishing or cross-correlation.

TripleBlind is far superior to sharing synthetic data because businesses can fully analyze real data in order to understand real trends. TripleBlind’s solution allows for data collaboration without jeopardizing privacy or compliance. Data shared through TripleBlind’s solution remains de-identified, private and can only be used for its intended purpose.

As shown in the above chart, collaboration via synthetic data has a negligible impact in most categories where accuracy and compliance are necessary. On the contrary, TripleBlind’s solution fulfills criteria across the board, making it a superior way to share data.

To learn more about how TripleBlind compares to other competitors and methods of data collaborations, follow us on LinkedIn and Twitter to be notified when we post the next installation in our Competitor Blog Series.

If you’d like to schedule a call or free demo to explore how TripleBlind can work for your business, please reach out to contact@tripleblind.ai

 

Read other blogs in this series:

Business Agreements
Homomorphic Encryption
Blockchain
Tokenization, Masking and Hashing
Federated Learning
Differential Privacy

How TripleBlind’s Solution Can Make Data Sharing in Healthcare More Horizontal

TripleBlind recently hosted a virtual roundtable discussion featuring thought leaders from Mayo Clinic and Novartis to explore the current state of data sharing in healthcare. TripleBlind’s co-founder and CEO, Riddhiman Das, was joined by Mayo Clinic’s Dr. Paul Friedman and Dr. Suraj Kapa and Sukant Mittal from Novartis. 


Current issues surrounding data sharing in healthcare

While the expansion of electronic medical records and technological advancements have led to vast amounts of health data, this data is not broadly shared due to concerns about personal identifiable information (PII) and protected health information (PHI).

When this data is not readily available to share and use, healthcare professionals cannot access information which would create a more equitable pool of patient data and lead to advancements in diagnosis and treatments. Doctors need a way to both respect patient privacy, but gain access to more comprehensive health histories. 


How the issue is currently being addressed

While complying with data privacy regulations, healthcare organizations are still doing all they can to ensure data pools are unbiased. 

Mayo Clinic currently validates independent cross populations – different ethnicities, races, etc. – within its own data sets. This task becomes more difficult when talking about a global population and the regulations that differ between different countries. 

Training data is essential to Mayo as they capture data from the broadest possible population. Mayo’s neural networks can detect subtle, interrelated patterns that translate the hidden signals the human body gives off all the time, but will not function properly if untrained. Today, Mayo has roughly 30 hospitals from four continents providing data and they are continually expanding as permitted.


In a perfect world, data sharing would be more horizontal

While institutions like Mayo work to remain unbiased and ethical, there remains a void globally across the healthcare industry to ethically and compliantly crowdsource patient information.

During the webinar with TripleBlind, Dr. Suraj Kapa mentioned that ideally, in the future of digital health, institutions could move away from monopolies of data and sharing data would be more horizontal. Organizations would be able to access data that reflects the broader concept of the world’s population rather than segmented, narrow cohorts of patients.

Compliantly sharing crowdsourced healthcare information in real time would create limitless possibilities and accelerate discovery and understanding for healthcare providers.


How TripleBlind can help healthcare institutions achieve this desired outcome

When it comes to private healthcare data, TripleBlind aims to enable the liquidity of this data in order to enable and foster innovation in healthcare.

TripleBlind’s groundbreaking solution allows highly-regulated enterprises like healthcare institutions to gain and share de-identified data without ever decrypting it. When de-identified data is shared, there is no chance of compliance issues or of the data being re-identified. TripleBlind enables institutions to leverage third-party data or allow third parties to use their data while guaranteeing that the data is going to be used for the stated purpose.

With TripleBlind’s technology, organizations can cover global ground rather than operating against the specific, narrow regulations that vary worldwide.

 


To learn more about how TripleBlind’s technology can open the door to compliant data sharing for your organization, please reach out to contact@tripleblind.com for a free demo. To watch a video of the roundtable featuring TripleBlind, Mayo Clinic and Novartis, visit here

How TripleBlind’s Data Privacy Solution Compares to Homomorphic Encryption

Homomorphic encryption is a technique that allows for computations to be done on encrypted data without needing a secret decryption key, allowing only the owner or those with the secret key to see the results of the computations. There are multiple applications in which fully homomorphic encryption can be applied, from something as simple as keeping a person’s Internet search history private from third-party marketers to more complicated computations such as those done with healthcare data. Homomorphic encryption is considered one of the more well-rounded encryption solutions in the market and has been adopted by tech giants like IBM and Microsoft.

However, homomorphic encryption’s most significant barrier to widespread use is how slow it is. In fact, according to IBM’s homomorphic encryption trials, it requires more than 42-times compute power and 20-times memory compared to other types of encryptions.

Homomorphic encryption’s speed is not the only place it falls short compared to TripleBlind’s data privacy technology. Below is a comparison chart of the two solutions:

  • TripleBlind
  • Fast
  • Universal, cloud based
  • Future proof
  • Blind interference supports all non-linear operations, including comparisons
  • Requires all parties online
  • All parties consent to each use
  • Mathematical digital rights management
  • Homomorphic Encryption
  • Slow
  • High CPU needs
  • May be cracked in the future
  • Only supports basic algebraic operations
  • Operates offline
  • Doesn’t require consent of all parties for other uses
  • No digital rights management

There are other areas in which homomorphic encryption doesn’t stack up compared to TripleBlind, including:

As you can see in the above charts, homomorphic encryption falls short in too many categories to provide an enterprise with a complete solution. Enterprises would likely need one or more other solutions to have all the criteria fulfilled.

Unlocking private data sharing with TripleBlind’s solution allows businesses to collaborate more fully, compliantly and across broader horizons than homomorphic encryption. To learn more about how TripleBlind compares to other competitors and methods of data collaborations, follow us on LinkedIn and Twitter to be notified when we post the next installation in our Competitor Blog Series.

If you’d like to schedule a call or free demo to explore how TripleBlind can work for your business, please reach out to contact@tripleblind.ai.

How TripleBlind’s Data Privacy Solution Compares to Traditional Business Agreements

Business agreements are a method of collaboration in which businesses share data under a predetermined set of rules, limitations and perimeters established between the parties. 

In a business agreement, “Business 1” might give “Business 2” access to their bank statements to reference purchases made throughout August 2021. Business 2 is expected only to reference purchases made in August 2021 and not access any other transactions or purchases made at any other time. And while Business 2 may be contractually obligated to stay within the agreement’s limitations, the raw data may still be made available and can be misused.

This data-sharing method requires trust between all parties involved because there is no way to regulate how the complete set of data will be used or with whom it will be shared. The level of trust that business agreements require takes time and resources to build and maintain, which creates limitations surrounding who businesses can collaborate with.

TripleBlind’s solution is superior to business agreements because it operates at zero-trust, increasing collaboration possibilities and improving efficiency in data partnerships. TripleBlind ensures that sensitive data is used for its intended purpose and cannot be abused.

TripleBlind keeps data and algorithms private at every stage of a data project and never stores data itself. Blind de-identification protects assets from various forms of misuse, including: 

  1. Malicious attempts to gain access to data or algorithms
  2. Trusted but curious parties viewing raw data
  3. Unauthorized uses of data resulting from raw data being left behind after a job is completed
  4. Violations of data privacy laws, including GDPR, HIPAA, and data residency, because the raw data is never moved from behind the firewall of the data owner.

Unlocking private data sharing with TripleBlind’s solution allows businesses to collaborate more fully, compliantly and across broader horizons than business agreements. To learn more about how TripleBlind compares to other competitors and methods of data collaborations, follow us on LinkedIn and Twitter to be notified when we post the next installation in our Competitor Blog Series.

If you’d like to schedule a call or free demo to explore how TripleBlind can work for your business, please reach out to contact@tripleblind.ai.

Read the other blogs in this series:
Homomorphic Encryption
Synthetic Data
Blockchain
Tokenization, Masking and Hashing
Federated Learning
Differential Privacy

TripleBlind Versus Competing Solutions

One of the most frequent questions we hear is, “How is TripleBlind different from other solutions?” Our technology is too detailed to explain in a short answer, especially when other technologies in this space are being developed by big names like Microsoft and IBM. However, it’s too important to go unaddressed. Over the course of several blogs, we will be going into detail about these other technologies such as homomorphic encryption, federated learning, blockchain, differential privacy, synthetic data, tokenization, and the old-fashioned – business agreement. 

Current solutions use one or some of these approaches, but TripleBlind is superior. We know that key providers such as Enveil, Baffle, Duality Technologies, Google, IBM, Intel, Microsoft, PreVeil offer homomorphic encryption and encryption-in-use camp. Some of the providers that leverage differential privacy include Immuta, Microsoft and SAP.

Providers that use synthetic data are Mostly.AI, Statice, Syntho and Tonic. For tokenization, IBM dominates with its Guardium and Cloud Pak for security offerings, as well as Informatica with its Informatica Data Privacy Management, making for a much larger, established market. 

Our mission is to help enterprises unlock the more than 90% of their data that goes unused due to data privacy and regulatory concerns. We change the game from “don’t be evil,” to “can’t be evil.” Arming enterprises with the ability to share and collaborate with that data creates opportunities that range from accelerating the creation and improving the accuracy of medical diagnoses to thwarting hackers and preventing the next big data breach. 

Today’s data privacy solutions are simply ineffective: business agreements hashed out by expensive lawyers take too long to negotiate and requires reliance on goodwill. Homomorphic encryption is slow, while secure enclaves are siloed. Masking and hashing a particular data element(s) reduces its accuracy. 

Differential privacy presents IP vulnerability, blockchain isn’t known to be future-proof, and federated learning has limited use for algorithms. Lastly, why use synthetic data when we could use real data for better results? 

TripleBlind enables enterprises to enforce compliance with any and all data privacy standards today – GDPR, HIPAA, PDPA and the myriad of state regulations popping up in the U.S. and data residency requirements in Asia. We also believe it will keep organizations in compliance with any future standards since its core architecture lets data providers share information with data users with data always remaining behind the provider’s firewall and all operations taking place behind the data user’s firewall.

We will dig deeper into the faults of these approaches and how they compare to TripleBlind. We’ll be sure to announce new blogs on our social media, so follow us on LinkedIn and Twitter. If you’re eager to learn more, schedule a call or demo for all the details at contact@tripleblind.ai.

Read the other blogs in this series:
Business Agreements
Homomorphic Encryption
Synthetic Data
Blockchain
Tokenization, Masking and Hashing
Federated Learning
Differential Privacy

Secure and Private Compute Summit