Alternatives to Rocket DataEdge

Compare Rocket DataEdge alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to Rocket DataEdge in 2026. Compare features, ratings, user reviews, pricing, and more from Rocket DataEdge competitors and alternatives in order to make an informed decision for your business.

  • 1
    dbt

    dbt

    dbt Labs

    dbt helps data teams transform raw data into trusted, analysis-ready datasets faster. With dbt, data analysts and data engineers can collaborate on version-controlled SQL models, enforce testing and documentation standards, lean on detailed metadata to troubleshoot and optimize pipelines, and deploy transformations reliably at scale. Built on modern software engineering best practices, dbt brings transparency and governance to every step of the data transformation workflow. Thousands of companies, from startups to Fortune 500 enterprises, rely on dbt to improve data quality and trust as well as drive efficiencies and reduce costs as they deliver AI-ready data across their organization. Whether you’re scaling data operations or just getting started, dbt empowers your team to move from raw data to actionable analytics with confidence.
    Compare vs. Rocket DataEdge View Software
    Visit Website
  • 2
    AnalyticsCreator

    AnalyticsCreator

    AnalyticsCreator

    AnalyticsCreator is a metadata-driven data warehouse automation application for teams working in the Microsoft data ecosystem. It enables data engineers to design, generate, and maintain production-ready data products across Microsoft SQL Server, Azure Data Factory, and Microsoft Fabric. By using centralized metadata, AnalyticsCreator generates ELT pipelines, dimensional models, historization logic, and analytical models in a consistent, version-controlled way. This reduces manual implementation effort and tool sprawl while ensuring transparency through built-in lineage tracking and clear visibility into data dependencies and change impact. With CI/CD integration via Azure DevOps and GitHub, plus support for custom SQL, AnalyticsCreator helps data teams scale delivery, enforce standards, and maintain control as complexity grows.
    Compare vs. Rocket DataEdge View Software
    Visit Website
  • 3
    IRI Voracity

    IRI Voracity

    IRI, The CoSort Company

    Voracity is the only high-performance, all-in-one data management platform accelerating AND consolidating the key activities of data discovery, integration, migration, governance, and analytics. Voracity helps you control your data in every stage of the lifecycle, and extract maximum value from it. Only in Voracity can you: 1) CLASSIFY, profile and diagram enterprise data sources 2) Speed or LEAVE legacy sort and ETL tools 3) MIGRATE data to modernize and WRANGLE data to analyze 4) FIND PII everywhere and consistently MASK it for referential integrity 5) Score re-ID risk and ANONYMIZE quasi-identifiers 6) Create and manage DB subsets or intelligently synthesize TEST data 7) Package, protect and provision BIG data 8) Validate, scrub, enrich and unify data to improve its QUALITY 9) Manage metadata and MASTER data. Use Voracity to comply with data privacy laws, de-muck and govern the data lake, improve the reliability of your analytics, and create safe, smart test data
  • 4
    AWS Glue

    AWS Glue

    Amazon

    AWS Glue is a serverless data integration service that makes it easy to discover, prepare, and combine data for analytics, machine learning, and application development. AWS Glue provides all the capabilities needed for data integration so that you can start analyzing your data and putting it to use in minutes instead of months. Data integration is the process of preparing and combining data for analytics, machine learning, and application development. It involves multiple tasks, such as discovering and extracting data from various sources; enriching, cleaning, normalizing, and combining data; and loading and organizing data in databases, data warehouses, and data lakes. These tasks are often handled by different types of users that each use different products. AWS Glue runs in a serverless environment. There is no infrastructure to manage, and AWS Glue provisions, configures, and scales the resources required to run your data integration jobs.
  • 5
    Rocket Data Virtualization
    Rocket® Data Virtualization (RDV) provides real‑time, zero‑copy access to mainframe and distributed data through a unified SQL abstraction layer. It lets organizations query and combine heterogeneous sources without data movement or replication, supporting analytics, AI/ML, and cloud‑native workloads while maintaining system‑of‑record governance. RDV connects to Db2, VSAM, IMS, Adabas, Datacom, and relational or cloud stores via JDBC/ODBC. Optimizations such as predicate pushdown, parallel execution, and adaptive caching deliver low‑latency access with minimal production impact. Security includes encryption, row/column‑level controls, data masking, and SAF/RACF integration, with high‑availability options for continuity. By keeping data in place and offloading read‑heavy workloads, RDV reduces mainframe CPU usage and enables governed enterprise data access without ETL or replication.
  • 6
    Rocket Data Replicate & Sync
    Rocket® Data Replicate and Sync (RDRS) provides high performance, sub-second change data capture (CDC) and bidirectional replication across mainframe, distributed, and cloud environments. It supports a broad set of enterprise sources and modern cloud targets, automatically capturing and applying data changes while adapting formats for each destination. The platform incorporates enterprise-grade encryption, secure connectivity such as site-to-site VPN and SAF controls, and built-in recovery and rollback to maintain data integrity across hybrid environments. By offloading data transformations to cloud compute, RDRS helps reduce mainframe CPU consumption while ensuring analytics, AI/ML, and operational applications receive current data. Supported sources include Db2, VSAM, IMS, Adabas, and Datacom, with delivery to targets such as Kafka, Snowflake, AWS, and Azure—without custom code or operational disruption.
  • 7
    Rocket Data Intelligence
    Rocket® Data Intelligence (RDI) delivers comprehensive visibility into enterprise data across mainframe, distributed, and cloud environments. It automatically discovers metadata, lineage, and data relationships so organizations can see where critical data resides, how it moves, and which applications and processes rely on it. RDI supports legacy and modern platforms, including Db2, VSAM, IMS, Adabas, Datacom, relational databases, ETL tools like Informatica and DataStage, code such as COBOL, Python, and Java, and cloud data stores. RDI provides enterprise-grade capabilities including automated data discovery and code parsing, impact analysis, lineage filtering, role/LOB-based categorization and governance, workflow management, business glossary, and dependency mapping. By unifying data asset visibility across hybrid environments, RDI reduces operational risk and accelerates data modernization, compliance reporting, discovery, and rationalization initiatives.
  • 8
    Denodo

    Denodo

    Denodo Technologies

    The core technology to enable modern data integration and data management solutions. Quickly connect disparate structured and unstructured sources. Catalog your entire data ecosystem. Data stays in the sources and it is accessed on demand, with no need to create another copy. Build data models that suit the needs of the consumer, even across multiple sources. Hide the complexity of your back-end technologies from the end users. The virtual model can be secured and consumed using standard SQL and other formats like REST, SOAP and OData. Easy access to all types of data. Full data integration and data modeling capabilities. Active Data Catalog and self-service capabilities for data & metadata discovery and data preparation. Full data security and data governance capabilities. Fast intelligent execution of data queries. Real-time data delivery in any format. Ability to create data marketplaces. Decoupling of business applications from data systems to facilitate data-driven strategies.
  • 9
    IBM Cloud Pak for Data
    The biggest challenge to scaling AI-powered decision-making is unused data. IBM Cloud Pak® for Data is a unified platform that delivers a data fabric to connect and access siloed data on-premises or across multiple clouds without moving it. Simplify access to data by automatically discovering and curating it to deliver actionable knowledge assets to your users, while automating policy enforcement to safeguard use. Further accelerate insights with an integrated modern cloud data warehouse. Universally safeguard data usage with privacy and usage policy enforcement across all data. Use a modern, high-performance cloud data warehouse to achieve faster insights. Empower data scientists, developers and analysts with an integrated experience to build, deploy and manage trustworthy AI models on any cloud. Supercharge analytics with Netezza, a high-performance data warehouse.
    Starting Price: $699 per month
  • 10
    Data Virtuality

    Data Virtuality

    Data Virtuality

    Connect and centralize data. Transform your existing data landscape into a flexible data powerhouse. Data Virtuality is a data integration platform for instant data access, easy data centralization and data governance. Our Logical Data Warehouse solution combines data virtualization and materialization for the highest possible performance. Build your single source of data truth with a virtual layer on top of your existing data environment for high data quality, data governance, and fast time-to-market. Hosted in the cloud or on-premises. Data Virtuality has 3 modules: Pipes, Pipes Professional, and Logical Data Warehouse. Cut down your development time by up to 80%. Access any data in minutes and automate data workflows using SQL. Use Rapid BI Prototyping for significantly faster time-to-market. Ensure data quality for accurate, complete, and consistent data. Use metadata repositories to improve master data management.
  • 11
    K2View

    K2View

    K2View

    At K2View, we believe that every enterprise should be able to leverage its data to become as disruptive and agile as the best companies in its industry. We make this possible through our patented Data Product Platform, which creates and manages a complete and compliant dataset for every business entity – on demand, and in real time. The dataset is always in sync with its underlying sources, adapts to changes in the source structures, and is instantly accessible to any authorized data consumer. Data Product Platform fuels many operational use cases, including customer 360, data masking and tokenization, test data management, data migration, legacy application modernization, data pipelining and more – to deliver business outcomes in less than half the time, and at half the cost, of any other alternative. The platform inherently supports modern data architectures – data mesh, data fabric, and data hub – and deploys in cloud, on-premise, or hybrid environments.
  • 12
    ETL DataHub
    DataHub from ETL Solutions is an enterprise-grade data integration, orchestration, and management platform designed to help organizations connect, harmonize, and operationalize data from diverse sources into a unified, governed, and accessible ecosystem. It enables seamless ingestion and transformation of structured and unstructured data through pre-built connectors and mappings, automated workflows, change data capture, and real-time data pipelines that support analytics, reporting, and AI/ML use cases. Built for hybrid and multi-cloud environments, DataHub centralizes metadata and business logic while enforcing data governance, lineage, and quality controls so stakeholders can trust and act on enterprise data. Its orchestration engine handles complex dependencies and schedules, ensuring data arrives on time and maintains consistency across systems.
  • 13
    CData Query Federation Drivers
    The Query Federation Drivers provide a universal data access layer that simplifies application development and data access. The drivers make it easy to query data across systems with SQL through a common driver interface. The Query Federation Drivers enable users to embed Logical Data Warehousing capabilities into any application or process. A Logical Data Warehouse is an architectural layer that enables access to multiple data sources on-demand, without relocating or transforming data in advance. Essentially the Query Federation Drivers give users simple, SQL-based access to all of your databases, data warehouses, and cloud applications through a single interface. Developers can pick multiple data processing systems and access all of them with a single SQL-based interface.
  • 14
    IBM DataStage
    Accelerate AI innovation with cloud-native data integration on IBM Cloud Pak for data. AI-powered data integration, anywhere. Your AI and analytics are only as good as the data that fuels them. With a modern container-based architecture, IBM® DataStage® for IBM Cloud Pak® for Data delivers that high-quality data. It combines industry-leading data integration with DataOps, governance and analytics on a single data and AI platform. Automation accelerates administrative tasks to help reduce TCO. AI-based design accelerators and out-of-the-box integration with DataOps and data science services speed AI innovation. Parallelism and multicloud integration let you deliver trusted data at scale across hybrid or multicloud environments. Manage the data and analytics lifecycle on the IBM Cloud Pak for Data platform. Services include data science, event messaging, data virtualization and data warehousing. Parallel engine and automated load balancing.
  • 15
    Fraxses

    Fraxses

    Intenda

    There are many products on the market that can help companies to do this, but if your priorities are to create a data-driven enterprise and to be as efficient and cost-effective as possible, then there is only one solution you should consider: Fraxses, the world’s foremost distributed data platform. Fraxses provides customers with access to data on demand, delivering powerful insights via a solution that enables a data mesh or data fabric architecture. Think of a data mesh as a structure that can be laid over disparate data sources, connecting them, and enabling them to function as a single environment. Unlike other data integration and virtualization platforms, the Fraxses data platform has a decentralized architecture. While Fraxses fully supports traditional data integration processes, the future lies in a new approach, whereby data is served directly to users without the need for a centrally owned data lake or platform.
  • 16
    Informatica PowerCenter
    Embrace agility with the market-leading scalable, high-performance enterprise data integration platform. Support the entire data integration lifecycle, from jumpstarting the first project to ensuring successful mission-critical enterprise deployments. PowerCenter, the metadata-driven data integration platform, jumpstarts and accelerates data integration projects in order to deliver data to the business more quickly than manual hand coding. Developers and analysts collaborate, rapidly prototype, iterate, analyze, validate, and deploy projects in days instead of months. PowerCenter serves as the foundation for your data integration investments. Use machine learning to efficiently monitor and manage your PowerCenter deployments across domains and locations.
  • 17
    CONNX

    CONNX

    Software AG

    Unlock the value of your data—wherever it resides. To become data-driven, you need to leverage all the information in your enterprise across apps, clouds and systems. With the CONNX data integration solution, you can easily access, virtualize and move your data—wherever it is, however it’s structured—without changing your core systems. Get your information where it needs to be to better serve your organization, customers, partners and suppliers. Connect and transform legacy data sources from transactional databases to big data or data warehouses such as Hadoop®, AWS and Azure®. Or move legacy to the cloud for scalability, such as MySQL to Microsoft® Azure® SQL Database, SQL Server® to Amazon REDSHIFT®, or OpenVMS® Rdb to Teradata®.
  • 18
    Palantir Foundry

    Palantir Foundry

    Palantir Technologies

    Foundry is a transformative data platform built to help solve the modern enterprise’s most critical problems by creating a central operating system for an organization’s data, while securely integrating siloed data sources into a common analytics and operations picture. Palantir works with commercial companies and government organizations alike to close the operational loop, feeding real-time data into your data science models and updating source systems. With a breadth of industry-leading capabilities, Palantir can help enterprises traverse and operationalize data to enable and scale decision-making, alongside best-in-class security, data protection, and governance. Foundry was named by Forrester as a leader in the The Forrester Wave™: AI/ML Platforms, Q3 2022. Scoring the highest marks possible in product vision, performance, market approach, and applications criteria. As a Dresner-Award winning platform, Foundry is the overall leader in the BI and Analytics market and rate
  • 19
    Enterprise Enabler

    Enterprise Enabler

    Stone Bond Technologies

    It unifies information across silos and scattered data for visibility across multiple sources in a single environment; whether in the cloud, spread across siloed databases, on instruments, in Big Data stores, or within various spreadsheets/documents, Enterprise Enabler can integrate all your data so you can make informed business decisions in real-time. By creating logical views of data from the original source locations. This means you can reuse, configure, test, deploy, and monitor all your data in a single integrated environment. Analyze your business data in one place as it is occurring to maximize the use of assets, minimize costs, and improve/refine your business processes. Our implementation time to market value is 50-90% faster. We get your sources connected and running so you can start making business decisions based on real-time data.
  • 20
    IBM InfoSphere Information Server
    Set up cloud environments quickly for ad hoc development, testing and productivity for your IT and business users. Reduce the risks and costs of maintaining your data lake by implementing comprehensive data governance, including end-to-end data lineage, for business users. Improve cost savings by delivering clean, consistent and timely information for your data lakes, data warehouses or big data projects, while consolidating applications and retiring outdated databases. Take advantage of automatic schema propagation to speed up job generation, type-ahead search, and backwards capability, while designing once and executing anywhere. Create data integration flows and enforce data governance and quality rules with a cognitive design that recognizes and suggests usage patterns. Improve visibility and information governance by enabling complete, authoritative views of information with proof of lineage and quality.
    Starting Price: $16,500 per month
  • 21
    Delphix

    Delphix

    Perforce

    Delphix is the industry leader in DataOps and provides an intelligent data platform that accelerates digital transformation for leading companies around the world. The Delphix DataOps Platform supports a broad spectrum of systems, from mainframes to Oracle databases, ERP applications, and Kubernetes containers. Delphix supports a comprehensive range of data operations to enable modern CI/CD workflows and automates data compliance for privacy regulations, including GDPR, CCPA, and the New York Privacy Act. In addition, Delphix helps companies sync data from private to public clouds, accelerating cloud migrations, customer experience transformation, and the adoption of disruptive AI technologies. Automate data for fast, quality software releases, cloud adoption, and legacy modernization. Source data from mainframe to cloud-native apps across SaaS, private, and public clouds.
  • 22
    TIBCO Platform

    TIBCO Platform

    Cloud Software Group

    TIBCO delivers industrial-strength solutions that meet your performance, throughput, reliability, and scalability needs while offering a wide range of technology and deployment options to deliver real-time data where it’s needed most. The TIBCO Platform will bring together an evolving set of your TIBCO solutions wherever they are hosted—in the cloud, on-premises, and at the edge—into a single, unified experience so that you can more easily manage and monitor them. TIBCO helps build solutions that are essential to the success of the world’s largest enterprises.
  • 23
    Algoreus

    Algoreus

    Turium AI

    All your data needs are delivered in one powerful platform. From data ingestion/integration, transformation, and storage to knowledge catalog, graph networks, data analytics, governance, monitoring, and, sharing. ​ An AI/ML platform that lets enterprises, train, test, troubleshoot, deploy, and govern models at scale to boost productivity while maintaining model performance in production with confidence. A dedicated solution for training models with minimal effort through AutoML or training your case-specific models from scratch with CustomML. Giving you the power to connect essential logic from ML with data. An integrated exploration of possible actions.​ Integration with your protocols and authorization models​. Propagation by default; extreme configurability at your service​. Leverage internal lineage system, for alerting and impact analysis​. Interwoven with the security paradigm; provides immutable tracking​.
  • 24
    Oracle Data Service Integrator
    Oracle Data Service Integrator provides companies the ability to quickly develop and manage federated data services for accessing single views of disparate information. Oracle Data Service Integrator is completely standards-based, declarative, and enables re-usability of data services. Oracle Data Service Integrator is the only data federation technology that supports the creation of bidirectional (read and write) data services from multiple data sources. In addition, Oracle Data Service Integrator offers the breakthrough capability of eliminating coding by graphically modeling both simple and complex updates to heterogeneous data sources. Install, verify, uninstall, upgrade, and get started with Data Service Integrator. Oracle Data Service Integrator was originally known as Liquid Data and AquaLogic Data Services Platform (ALDSP). Some instances of the original names remain in the product, installation path, and components.
  • 25
    TIBCO Data Virtualization
    An enterprise data virtualization solution that orchestrates access to multiple and varied data sources and delivers the datasets and IT-curated data services foundation for nearly any solution. As a modern data layer, the TIBCO® Data Virtualization system addresses the evolving needs of companies with maturing architectures. Remove bottlenecks and enable consistency and reuse by providing all data, on demand, in a single logical layer that is governed, secure, and serves a diverse community of users. Immediate access to all data helps you develop actionable insights and act on them in real time. Users are empowered because they can easily search for and select from a self-service directory of virtualized business data and then use their favorite analytics tools to obtain results. They can spend more time analyzing data, less time searching for it.
  • 26
    Timbr.ai

    Timbr.ai

    Timbr.ai

    Timbr is the ontology-based semantic layer used by leading enterprises to make faster, better decisions with ontologies that transform structured data into AI-ready knowledge. By unifying enterprise data into a SQL-queryable knowledge graph, Timbr makes relationships, metrics, and context explicit, enabling both humans and AI to reason over data with accuracy and speed. Its open, modular architecture connects directly to existing data sources, virtualizing and governing them without replication. The result is a dynamic, easily accessible model that powers analytics, automation, and LLMs through SQL, APIs, SDKs, and natural language. Timbr lets organizations operationalize AI on their data - securely, transparently, and without dependence on proprietary stacks - maximizing data ROI and enabling teams to focus on solving problems instead of managing complexity.
    Starting Price: $599/month
  • 27
    Varada

    Varada

    Varada

    Varada’s dynamic and adaptive big data indexing solution enables to balance performance and cost with zero data-ops. Varada’s unique big data indexing technology serves as a smart acceleration layer on your data lake, which remains the single source of truth, and runs in the customer cloud environment (VPC). Varada enables data teams to democratize data by operationalizing the entire data lake while ensuring interactive performance, without the need to move data, model or manually optimize. Our secret sauce is our ability to automatically and dynamically index relevant data, at the structure and granularity of the source. Varada enables any query to meet continuously evolving performance and concurrency requirements for users and analytics API calls, while keeping costs predictable and under control. The platform seamlessly chooses which queries to accelerate and which data to index. Varada elastically adjusts the cluster to meet demand and optimize cost and performance.
  • 28
    DataVantage

    DataVantage

    DataVantage

    ​DataVantage offers comprehensive data management solutions designed to secure and manage sensitive information across mainframe and distributed environments. Key offerings include DataVantage for IMS, Db2, and VSAM, which provide advanced data masking, editing, and extraction capabilities to protect Personally Identifiable Information (PII) during non-production processes. DataVantage DME (Data Masking Express) offers cost-effective, on-the-fly data masking for Db2, IMS, and VSAM environments, ensuring compliance without disrupting existing workflows. For distributed systems, DataVantage Global delivers robust data masking, obfuscation, and de-identification, supporting compliance and operational efficiency across platforms. Additionally, DataVantage Adviser simplifies COBOL file management post-mainframe rehosting or application modernization, enhancing data accessibility and editing.
  • 29
    Harbr

    Harbr

    Harbr

    Create data products from any source in seconds, without moving the data. Make them available to anyone, while maintaining complete control. Deliver powerful experiences to unlock value. Enhance your data mesh by seamlessly sharing, discovering, and governing data across domains. Foster collaboration and accelerate innovation with unified access to high-quality data products. Provide governed access to AI models for any user. Control how data interacts with AI to safeguard intellectual property. Automate AI workflows to rapidly integrate and iterate new capabilities. Access and build data products from Snowflake without moving any data. Experience the ease of getting more from your data. Make it easy for anyone to analyze data and remove the need for centralized provisioning of infrastructure and tools. Data products are magically integrated with tools, to ensure governance and accelerate outcomes.
  • 30
    Dremio

    Dremio

    Dremio

    Dremio delivers lightning-fast queries and a self-service semantic layer directly on your data lake storage. No moving data to proprietary data warehouses, no cubes, no aggregation tables or extracts. Just flexibility and control for data architects, and self-service for data consumers. Dremio technologies like Data Reflections, Columnar Cloud Cache (C3) and Predictive Pipelining work alongside Apache Arrow to make queries on your data lake storage very, very fast. An abstraction layer enables IT to apply security and business meaning, while enabling analysts and data scientists to explore data and derive new virtual datasets. Dremio’s semantic layer is an integrated, searchable catalog that indexes all of your metadata, so business users can easily make sense of your data. Virtual datasets and spaces make up the semantic layer, and are all indexed and searchable.
  • 31
    SAS Data Management

    SAS Data Management

    SAS Institute

    No matter where your data is stored, from cloud, to legacy systems, to data lakes, like Hadoop, SAS Data Management helps you access the data you need. Create data management rules once and reuse them, giving you a standard, repeatable method for improving and integrating data, without additional cost. As an IT expert, it's easy to get entangled in tasks outside your normal duties. SAS Data Management enables your business users to update data, tweak processes and analyze results themselves, freeing you up for other projects. Plus, a built-in business glossary, as well as SAS and third-party metadata management and lineage visualization capabilities, keep everyone on the same page. SAS Data Management technology is truly integrated, which means you’re not forced to work with a solution that’s been cobbled together. All our components, from data quality to data federation technology, are part of the same architecture.
  • 32
    SAS Federation Server
    Create federated source data names to enable users to access multiple data sources via the same connection. Use the web-based administrative console for simplified maintenance of user access, privileges and authorizations. Apply data quality functions such as match-code generation, parsing and other tasks inside the view. Improved performance with in-memory data caches & scheduling. Secured information with data masking & encryption. Lets you keep application queries current and available to users, and reduce loads on operational systems. Enables you to define access permissions for a user or group at the catalog, schema, table, column and row levels. Advanced data masking and encryption capabilities let you determine not only who’s authorized to view your data, but also what they see on an extremely granular level. It all helps ensure sensitive data doesn’t fall into the wrong hands.
  • 33
    Precisely Connect
    Integrate data seamlessly from legacy systems into next-gen cloud and data platforms with one solution. Connect helps you take control of your data from mainframe to cloud. Integrate data through batch and real-time ingestion for advanced analytics, comprehensive machine learning and seamless data migration. Connect leverages the expertise Precisely has built over decades as a leader in mainframe sort and IBM i data availability and security to lead the industry in accessing and integrating complex data. Access to all your enterprise data for the most critical business projects is ensured by support for a wide range of sources and targets for all your ELT and CDC needs.
  • 34
    Sesame Software

    Sesame Software

    Sesame Software

    Sesame Software specializes in secure, efficient data integration and replication across diverse cloud, hybrid, and on-premise sources. Our patented scalability ensures comprehensive access to critical business data, facilitating a holistic view in the BI tools of your choice. This unified perspective empowers your own robust reporting and analytics, enabling your organization to regain control of your data with confidence. At Sesame Software, we understand what’s at stake when you need to move a massive amount of data between environments quickly—while keeping it protected, maintaining centralized access, and ensuring compliance with regulations. Over the past 30+ years, we’ve helped hundreds of organizations like Proctor & Gamble, Bank of America, and the U.S. government connect, move, store, and protect their data.
  • 35
    Oracle Big Data SQL Cloud Service
    Oracle Big Data SQL Cloud Service enables organizations to immediately analyze data across Apache Hadoop, NoSQL and Oracle Database leveraging their existing SQL skills, security policies and applications with extreme performance. From simplifying data science efforts to unlocking data lakes, Big Data SQL makes the benefits of Big Data available to the largest group of end users possible. Big Data SQL gives users a single location to catalog and secure data in Hadoop and NoSQL systems, Oracle Database. Seamless metadata integration and queries which join data from Oracle Database with data from Hadoop and NoSQL databases. Utilities and conversion routines support automatic mappings from metadata stored in HCatalog (or the Hive Metastore) to Oracle Tables. Enhanced access parameters give administrators the flexibility to control column mapping and data access behavior. Multiple cluster support enables one Oracle Database to query multiple Hadoop clusters and/or NoSQL systems.
  • 36
    Lyftrondata

    Lyftrondata

    Lyftrondata

    Whether you want to build a governed delta lake, data warehouse, or simply want to migrate from your traditional database to a modern cloud data warehouse, do it all with Lyftrondata. Simply create and manage all of your data workloads on one platform by automatically building your pipeline and warehouse. Analyze it instantly with ANSI SQL, BI/ML tools, and share it without worrying about writing any custom code. Boost the productivity of your data professionals and shorten your time to value. Define, categorize, and find all data sets in one place. Share these data sets with other experts with zero codings and drive data-driven insights. This data sharing ability is perfect for companies that want to store their data once, share it with other experts, and use it multiple times, now and in the future. Define dataset, apply SQL transformations or simply migrate your SQL data processing logic to any cloud data warehouse.
  • 37
    Oracle Big Data Preparation
    Oracle Big Data Preparation Cloud Service is a managed Platform as a Service (PaaS) cloud-based offering that enables you to rapidly ingest, repair, enrich, and publish large data sets with end-to-end visibility in an interactive environment. You can integrate your data with other Oracle Cloud Services, such as Oracle Business Intelligence Cloud Service, for down-stream analysis. Profile metrics and visualizations are important features of Oracle Big Data Preparation Cloud Service. When a data set is ingested, you have visual access to the profile results and summary of each column that was profiled, and the results of duplicate entity analysis completed on your entire data set. Visualize governance tasks on the service Home page with easily understood runtime metrics, data health reports, and alerts. Keep track of your transforms and ensure that files are processed correctly. See the entire data pipeline, from ingestion to enrichment and publishing.
  • 38
    Actifio

    Actifio

    Google

    Automate self-service provisioning and refresh of enterprise workloads, integrate with existing toolchain. High-performance data delivery and re-use for data scientists through a rich set of APIs and automation. Recover any data across any cloud from any point in time – at the same time – at scale, beyond legacy solutions. Minimize the business impact of ransomware / cyber attacks by recovering quickly with immutable backups. Unified platform to better protect, secure, retain, govern, or recover your data on-premises or in the cloud. Actifio’s patented software platform turns data silos into data pipelines. Virtual Data Pipeline (VDP) delivers full-stack data management — on-premises, hybrid or multi-cloud – from rich application integration, SLA-based orchestration, flexible data movement, and data immutability and security.
  • 39
    VeloX Software Suite

    VeloX Software Suite

    Bureau Of Innovative Projects

    VeloX Software Suite enables Data Migration and System Integration throughout the entire organization. The suite consists of two applications, Migration Studio (VXm) for user-controlled data migrations; Integration Server (VXi), for automated data processing and integration. Extract from multiple sources and propagate to multiple destinations. Near real-time unified view of data without moving between sources. Physically bring data together from a multitude of sources, reduce the number of data storage locations, and transform based on business rules. Extract from multiple sources and propagate to multiple destinations. Event- and rules-driven. Synchronous and asynchronous exchange. EAI and EDR technologies. Near real-time unified view of data without moving between sources. Service-oriented architecture. Various abstraction and transformation techniques. EII technologies.
  • 40
    ibi Omni-Gen

    ibi Omni-Gen

    Cloud Software Group

    The modern, highly scalable ibi Omni-Gen Data Integration Framework provides powerful data integration and cleansing technologies that ensure your data is timely, accurate, consistent, and accessible. Interoperable Omni-Gen architecture insulates end users from data complexities and ensures delivery of the right data to the right place at the right time for faster, smarter decisions. With Omni-Gen, you can more easily break down data silos and add new data sources, migrate legacy systems, manage M&A activities, and achieve better results from your digital transformation efforts. Rich data profiling capabilities easily detect anomalies and identify opportunities for further refinement of data quality, match/merge, and remediation rules. You can seamlessly integrate data from any source and structure into a commonly understood canonical model. Don't just enforce data governance, provide a complete, historical, browser view of all golden records mastered across functional domains.
  • 41
    Red Hat JBoss Data Virtualization
    Red Hat JBoss Data Virtualization is a lean, virtual data integration solution that unlocks trapped data and delivers it as easily consumable, unified, and actionable information. Red Hat JBoss Data Virtualization makes data spread across physically diverse systems, such as multiple databases, XML files, and Hadoop systems, appear as a set of tables in a local database. Provides standards-based read/write access to heterogeneous data stores in real-time. Speeds application development and integration by simplifying access to distributed data. Integrate and transform data semantics based on data consumer requirements. Provides centralized access control, and auditing through robust security infrastructure. Turn fragmented data into actionable information at the speed your business needs. Red Hat offers support and maintenance over stated time periods for the major versions of JBoss products.
  • 42
    Orbit Analytics

    Orbit Analytics

    Orbit Analytics

    Empower your business by leveraging a true self-service reporting and analytics platform. Powerful and scalable, Orbit’s operational reporting and business intelligence software enables users to create their own analytics and reports. Orbit Reporting + Analytics offers pre-built integration with enterprise resource planning (ERP) and key cloud business applications that include PeopleSoft, Oracle E-Business Suite, Salesforce, Taleo, and more. With Orbit, you can quickly and efficiently find answers from any data source, determine opportunities, and make smart, data-driven decisions. Orbit comes with more than 200 integrators and connectors that allow you to combine data from multiple data sources, so you can harness the power of collective knowledge to make informed decisions. Orbit Adapters connect with your key business systems, and designed to seamlessly inherit authentication, data security, business roles and apply them to reporting.
  • 43
    Qlik Replicate
    Qlik Replicate is a high-performance data replication tool offering optimized data ingestion from a broad array of data sources and platforms and seamless integration with all major big data analytics platforms. Replicate supports bulk replication as well as real-time incremental replication using CDC (change data capture). Our unique zero-footprint architecture eliminates unnecessary overhead on your mission-critical systems and facilitates zero-downtime data migrations and database upgrades. Database replication enables you to move or consolidate data from a production database to a newer version of the database, another type of computing environment, or an alternative database management system, to migrate data from SQL Server to Oracle, for example. Data replication can be used to offload production data from a database, and load it to operational data stores or data warehouses for reporting or analytics.
  • 44
    data.world

    data.world

    data.world

    data.world is a fully managed service, born in the cloud, and optimized for modern data architectures. That means we handle all updates, migrations, and maintenance. Set up is fast and simple with a large and growing ecosystem of pre-built integrations including all of the major cloud data warehouses. When time-to-value is critical, your team needs to solve real business problems, not fight with hard-to-manage data software. data.world makes it easy for everyone, not just the "data people", to get clear, accurate, fast answers to any business question. Our cloud-native data catalog maps your siloed, distributed data to familiar and consistent business concepts, creating a unified body of knowledge anyone can find, understand, and use. In addition to our enterprise product, data.world is home to the world’s largest collaborative open data community. It’s where people team up on everything from social bot detection to award-winning data journalism.
    Starting Price: $12 per month
  • 45
    Virtuoso

    Virtuoso

    OpenLink Software

    Virtuoso Universal Server is a modern platform built on existing open standards that harnesses the power of Hyperlinks ( functioning as Super Keys ) for breaking down data silos that impede both user and enterprise ability. Using Virtuoso, you can easily generate financial profile knowledge graphs from near real time financial activity that reduce the cost and complexity associated with detecting fraudent activity patterns. Courtesy of its high-performance, secure, and scalable dbms engine, you can use intelligent reasoning and inference to harmonize fragmented identities using personally identifying attributes such as email addresses, phone numbers, social-security numbers, drivers licenses, etc. for building fraud detection solutions. Virtuoso helps you build powerful solutions applications driven by knowledge graphs derived from a variety of life sciences oriented data sources.
    Starting Price: $42 per month
  • 46
    Rocket Secure Host Access
    Rocket® Secure Host Access is a security-first terminal emulation solution designed to protect access to critical host and mainframe applications. It extends modern identity and access management (IAM) controls directly to green screen environments. The platform enables phishing-resistant, passwordless authentication to reduce credential-based threats. Rocket Secure Host Access supports compliance with regulations such as HIPAA, PCI-DSS, and DORA. It works across desktop, web, and hybrid environments for flexible deployment and consolidated ongoing administration. Strong encryption standards like TLS 1.3 and SSH help safeguard sensitive data. The solution strengthens enterprise security without disrupting existing workflows.
  • 47
    Tokern

    Tokern

    Tokern

    Open source data governance suite for databases and data lakes. Tokern is a simple to use toolkit to collect, organize and analyze data lake's metadata. Run as a command-line app for quick tasks. Run as a service for continuous collection of metadata. Analyze lineage, access control and PII datasets using reporting dashboards or programmatically in Jupyter notebooks. Tokern is an open source data governance suite for databases and data lakes. Improve ROI of your data, comply with regulations like HIPAA, CCPA and GDPR and protect critical data from insider threats with confidence. Centralized metadata management of users, datasets and jobs. Powers other data governance features. Track Column Level Data Lineage for Snowflake, AWS Redshift and BigQuery. Build lineage from query history or ETL scripts. Explore lineage using interactive graphs or programmatically using APIs or SDKs.
  • 48
    CData Sync

    CData Sync

    CData Software

    CData Sync is a universal data pipeline that delivers automated continuous replication between hundreds of SaaS applications & cloud data sources and any major database or data warehouse, on-premise or in the cloud. Replicate data from hundreds of cloud data sources to popular database destinations, such as SQL Server, Redshift, S3, Snowflake, BigQuery, and more. Configuring replication is easy: login, select the data tables to replicate, and select a replication interval. Done. CData Sync extracts data iteratively, causing minimal impact on operational systems by only querying and updating data that has been added or changed since the last update. CData Sync offers the utmost flexibility across full and partial replication scenarios and ensures that critical data is stored safely in your database of choice. Download a 30-day free trial of the Sync application or request more information at www.cdata.com/sync
  • 49
    TapData

    TapData

    TapData

    CDC-based live data platform for heterogeneous database replication, real-time data integration, or building a real-time data warehouse. By using CDC to sync production line data stored in DB2 and Oracle to the modern database, TapData enabled an AI-augmented real-time dispatch software to optimize the semiconductor production line process. The real-time data made instant decision-making in the RTD software a possibility, leading to faster turnaround times and improved yield. As one of the largest telcos, customer has many regional systems that cater to the local customers. By syncing and aggregating data from various sources and locations into a centralized data store, customers were able to build an order center where the collective orders from many applications can now be aggregated. TapData seamlessly integrates inventory data from 500+ stores, providing real-time insights into stock levels and customer preferences, enhancing supply chain efficiency.
  • 50
    Ascend

    Ascend

    Ascend

    Ascend gives data teams a unified and automated platform to ingest, transform, and orchestrate their entire data engineering and analytics engineering workloads, 10X faster than ever before.​ Ascend helps gridlocked teams break through constraints to build, manage, and optimize the increasing number of data workloads required. Backed by DataAware intelligence, Ascend works continuously in the background to guarantee data integrity and optimize data workloads, reducing time spent on maintenance by up to 90%. Build, iterate on, and run data transformations easily with Ascend’s multi-language flex-code interface enabling the use of SQL, Python, Java, and, Scala interchangeably. Quickly view data lineage, data profiles, job and user logs, system health, and other critical workload metrics at a glance. Ascend delivers native connections to a growing library of common data sources with our Flex-Code data connectors.
    Starting Price: $0.98 per DFC