FlyingBinary Ltd

Open Source Big Data Analyse Platform

Featuring distributed, fault-tolerant architecture with true NoSQL database, enterprise search and analytics, Open Source Big Data Analyse Platform resides in FlyingBinary's open source and big data suite, simplifying data integration, management, analysis and visualisation. Cassandra database supports true 24x7. Solr search delivers Google-like experience, while Apache Spark delivers citizen-scale analytics.


  • Always-on distributed and redundant architecture
  • Fault tolerant from component to data centre level
  • Linear scalability to massive data and transaction volumes
  • Open source and commodity hardware
  • Enterprise support and security is built-in
  • Data architecture supports transaction and warehousing workloads
  • Complements Plotly Private Visualisation service
  • Simple, Google-like integrated search eliminates separate data store
  • Integrated analytics eliminates data proliferation
  • Web-scale analytics available on entire production database


  • Advanced architecture eliminates most downtime scenarios
  • Supports true 24x7 availability and operation
  • Scales to citizen-scale or web-scale workloads
  • Open and commodity components provide lowest cost of ownership
  • Enterprise support assures production integrity
  • Single database for both transaction and warehousing reduces cost
  • Integrated search reduces cost and simplifies management
  • Simple search interface minimises training costs
  • Integrated analytics supports continual access to latest data
  • Web-scale analytics removes need for separate marts or cubes


£24,600 a terabyte a year

Service documents


G-Cloud 12

Service ID

1 9 4 9 3 8 9 0 2 9 8 2 8 7 4


FlyingBinary Ltd Dr Jacqui Taylor
Telephone: +44 77 899 668 02

Service scope

Service constraints
The service is deployed only on Linux nodes. Accessing management and design functions (using DataStax Studio and OpsCenter) requires a 64-bit operating system (Windows 8 or 10, Linux or macOS) with Oracle Java 11 and Safari, Chrome or Firefox. Note that Internet Explorer and Microsoft Edge are not supported browsers.
System requirements
Windows, Linux or macOS clients to access management functions

User support

Email or online ticketing support
Email or online ticketing
Support response times
We respond to all questions within 8 business hours. Normal business hours are 0800-1800 Mon-Fri excluding UK holidays. We regularly respond to questions within 4 business hours, and we triage all inbound questions to establish urgency and set appropriate priorities.
User can manage status and priority of support tickets
Phone support
Phone support availability
9 to 5 (UK time), Monday to Friday
Web chat support
Onsite support
Yes, at extra cost
Support levels
Our standard support service is by email 0800-1800 Mon-Fri excluding UK holidays and we respond to all questions within 8 business hours. We regularly respond to questions within 4 business hours, and we triage all inbound questions to establish urgency and set appropriate priorities. We can offer several extensions to the standard service at extra cost: extended hours support at 10x7 or 24x7; 4 business hour response; phone support; web chat support; onsite support. The support service is provided for issues, incidents and service requests. We will also respond to questions about how to use the service, but we reserve the right to direct clients to our training services if users are clearly not equipped to use the service competently.
Support available to third parties

Onboarding and offboarding

Getting started
There are four types of onboarding help available, which may be used in combination: online help; key contact getting started assistance; additional onboarding assistance; service training courses. Key contact getting started assistance is included within the cost of the service. It is usually conducted as an interactive web screenshare session, but may also be provided as a pre-recorded video session with a follow up teleconference, or as an onsite session. Onsite sessions outside the M25 area will incur an additional cost. Online help is available to all client users of the service and is included within the cost. Additional onboarding assistance and service training courses are both provided via the companion Cloud Support service.
Service documentation
Documentation formats
End-of-contract data extraction
The service has built-in tools that allow user data extraction. These tools are available at any time, not just at contract end, and may be used for any data extraction purpose required. There are no restrictions: all user data may be extracted from the service for any reason, at any time. User data extraction does not require supplier intervention, but clients may choose to limit this functionality to client super users or administrators. There is no limit to the number or size of data extractions performed, but the service is priced on a "fair use" basis for network charges. Excessively large or very frequent data extractions may incur an additional cost or bandwidth limits. Alternatively, the companion Cloud Support service offers assistance with data extraction, at an additional cost.
End-of-contract process
Prior to the end of contract/contract renewal date, we discuss the forward plan with the client key contact. In the event of contract end, we work with the key contact to ensure that all user data and collateral is retrieved or extracted before the contract end date. Although there are standard data extraction tools built in to the service, there may be a requirement for special data migration to the client's chosen new service. In that instance, we can work with the new supplier to migrate data to maximum benefit for the client, such work would be performed within the companion Cloud Support service and incur an additional charge. In the event that the client is unable to extract all required data before the contract end date, it is often possible to extend the duration of the data extraction features of the service for an additional period; or to extend the full service for an additional period; so that the client does not suffer unnecessary unavailability.

Using the service

Web browser interface
Using the web interface
The Open Source Big Data suite of services is built upon the DataStax Enterprise product set and includes the built-in DataStax OpsCenter web-based visual management and monitoring solution. OpsCenter has integrated functionality for real-time monitoring, tuning, provisioning, backup and security management. The ring management aspects handled by OpsCenter, include node provisioning and repair, and ring reallocation. OpsCenter provides a consolidated view of monitoring metrics, and allows ring tuning parameters to be changed directly from the console. OpsCentre does not handle any direct data or metadata management or query functions.
Web interface accessibility standard
None or don’t know
How the web interface is accessible
The underlying framework in the web user interface is Dojo, which has made considerable efforts to be compliant with WCAG 2.0 A standards. The official statement is, “While Dojo does not make an explicit claim of compliance to W3C WCAG 2.0 Level A or to US Section 508, every attempt has been made to meet those guidelines and to make the 1.0 and future versions of the core widget set, dijit, accessible to keyboard, low vision and assistive technology users.”
Web interface accessibility testing
While the core web user interface framework has made considerable efforts to be compliant with WCAG 2.0 A standards, no official testing with metricated outputs has yet been completed.
What users can and can't do using the API
The API includes several language drivers allowing applications built in C/C++, C#, Java, Node.js, Python and Ruby. Practically everything can be automated through the API, from basic cluster operations such as auto node discovery and connection pooling, through to CRUD data manipulation, query execution and table management. The API supports synchronous and asynchronous (non-blocking) modes, with simple, prepared and batch statements. The fluent API includes a query builder and an object mapper that simplifies converting domain classes to and from query results. Convenient execution features include automatic reconnection, configurable load balancing and automatic configuration for any cluster size.
API automation tools
  • Ansible
  • Chef
  • OpenStack
  • Puppet
  • Other
Other API automation tools
  • Docker
  • Mesos
API documentation
API documentation formats
Command line interface
Command line interface compatibility
Linux or Unix
Using the command line interface
Several command line tools are provided, both for cluster management and data manipulation. Preflight tools verify cluster configuration; management tools provide node start and stop functions and cluster control. Other utility tools manage external clients and perform cluster stress testing. The Cassandra Query Language (CQL) shell allows users to issue data definition statements to manage data storage objects including keyspaces, tables and views. Users can also manipulate data directly with full CRUD capability, including support for user defined types and aggregates. Cassandra select statements allow users to retrieve data in raw form, or as the results from native or user-defined aggregates or transformation functions.


Scaling available
Scaling type
Independence of resources
Our testing tells us the lower and upper levels of cloud resource required to maintain performance for normal levels of demand. We reserve and dedicate the lower level of cloud resources to each client, so that no client user can impact the performance of any other client. As client demand increases, we monitor performance and ensure that additional cloud resources are available to maintain performance, to the upper level. The service is priced on a "fair use" basis, so exceeding the upper level of demand means either reduced performance, or request further cloud resources, at additional cost.
Usage notifications


Infrastructure or application metrics
Metrics types
  • Memory
  • Number of active instances
  • Other
Other metrics
  • Java Virtual Machine (JVM) heap
  • Operations per second
  • Node status
  • Latency
Reporting types
  • API access
  • Real-time dashboards
  • Regular reports
  • Reports on request


Supplier type
Not a reseller

Staff security

Staff security clearance
Conforms to BS7858:2012
Government security clearance
Up to Developed Vetting (DV)

Asset protection

Knowledge of data storage and processing locations
Data storage and processing locations
  • United Kingdom
  • European Economic Area (EEA)
User control over data storage and processing locations
Datacentre security standards
Complies with a recognised standard (for example CSA CCM version 3.0)
Penetration testing frequency
At least once a year
Penetration testing approach
Another external penetration testing organisation
Protecting data at rest
Physical access control, complying with SSAE-16 / ISAE 3402
Data sanitisation process
Data sanitisation type
Explicit overwriting of storage before reallocation
Equipment disposal approach
Complying with a recognised standard, for example CSA CCM v.30, CAS (Sanitisation) or ISO/IEC 27001

Backup and recovery

Backup and recovery

Data-in-transit protection

Data protection between buyer and supplier networks
TLS (version 1.2 or above)
Data protection within supplier network
TLS (version 1.2 or above)

Availability and resilience

Guaranteed availability
We guarantee the service will be available 99.5% of the time during business hours. Measured over the last two years, we have comfortably exceeded 99.9% availability. For use cases requiring higher availablity guarantees, we can discuss alternate cloud deployments and architectures, at additional cost. No financial recompense is offered for a single failure to meet service level within a year. In the event there are two or more failures in successive months, then a discount equal to one twelfth of the annual cost will be offered for renewals at the next anniversary.
Approach to resilience
All cloud resources are virtualised, so there is no dependency on any single hardware component. The underlying infrastructure is 99.999% available. As standard, all service components are deployed in a single availability zone to minimise service costs. All client deployments are managed as separate cloud instances, ensuring that performance-affecting activity occurring in one client deployment cannot impact any others. In addition to the single zone service, there is also the option of dual zone service, where all service components are replicated at a second site, at additional cost. The dual availability zone option provides geographic resilience, since all data is replicated across zones. It also provides additional throughput capacity, since the two zones operate in an active-active architecture. A single zone backup option is also available at additional cost. As standard, primary or single zone hosting is located in the UK, while secondary zone hosting is located in the EEA.
Outage reporting
Service affecting incidents are reported to client key contacts by email. Service affecting incidents include any observation of diminished service performance, reduced or missing functionality and not just observation of service outage. Initial emails may include only a report of the observation. Follow up emails will include notification of full service restoration, or estimated time to restoration, or additional details of mitigations or workarounds. Client key contacts may also request further details or additional help, as the emails are issued by a client support team, capable of responding to requests for additional information; the emails are not issued by "no-reply" mailboxes.

Identity and authentication

User authentication
  • Public key authentication (including by TLS client certificate)
  • Limited access network (for example PSN)
  • Username or password
Access restrictions in management interfaces and support channels
Opscenter is installed on a client machine, for example an admin's Linux or Mac laptop. It can also be installed on a dedicated server and accessed remotely. Depending on the choice of management machine, the access can be restricted to a separate, dedicated network access point. Within Opscenter, individual user accounts are assigned to one or more roles. The roles contain permissions for different access requirements, and therefore provide access on a specific grant basis.
Access restriction testing frequency
At least once a year
Management access authentication
  • Public key authentication (including by TLS client certificate)
  • Limited access network (for example PSN)
  • Username or password
Devices users manage the service through
  • Dedicated device on a segregated network (providers own provision)
  • Dedicated device on a government network (for example PSN)
  • Dedicated device over multiple services or networks
  • Any device but through a bastion host (a bastion host is a server that provides access to a private network from an external network such as the internet)
  • Directly from any device which may also be used for normal business (for example web browsing or viewing external email)

Audit information for users

Access to user activity audit information
Users have access to real-time audit information
How long user audit data is stored for
Access to supplier activity audit information
Users have access to real-time audit information
How long supplier audit data is stored for
How long system logs are stored for

Standards and certifications

ISO/IEC 27001 certification
ISO 28000:2007 certification
CSA STAR certification
PCI certification
Other security certifications
Any other security certifications
Cyber Essentials

Security governance

Named board-level person responsible for service security
Security governance certified
Security governance approach
Our security governance is meshed with our corporate governance, with our CEO having final responsibility for G-Cloud services and governance policies and our CTO having day-to-day responsibility for policy implementation. Our policies cover people, processes, technology and information assets, at board, management and operational levels. We review our policies and update our practices in light of new regulation, standards and best practice to ensure we are able to counter current and emerging threats.
Information security policies and processes
Our information security policy follows guidance in ISO27001, is owned by our CEO and implemented by our CTO. The top level Information Security Policy is supported by Architecture, Operations and Client Access policies. The policy states security objectives and establishes principles to ensure current and continued adherance and continual improvement. The policy set is integral to staff induction and all staff are required to agree and accept that information security governance is a core working principle. Operational checklists enforce security practices at the day to day level, and activities cannot be signed off without verified completion. The checklists also contain sections for feedback and challenge so that we actively improve. All staff are expected to challenge, because if we don't, bad actors will. Standard reporting flows from operational analyst or team lead to manager to CTO, but any staff member can invoke exceptional reporting directly to management or board, to ensure that important issues receive appropriate attention. Internal reviews ensure that all information security processes are working smoothly and as designed. If exceptions are found, we perform root cause analysis to understand if/how we need to change working practice to support our information security objectives.

Operational security

Configuration and change management standard
Supplier-defined controls
Configuration and change management approach
Service hardware components are managed by our data centre suppliers. Component lifecycle management is accredited to ISO27001, including storage overwrite and secure destruction. Service software is assembled from existing software components (proprietary and open source), is obtained only from the official repository, and security checked before use. No custom software is used. The service deployment checklist records all versions and change dates. Planned infrastructure or software changes are reviewed for new or changed features or capabilities, and internal software library dependencies. If needed, configurations are changed to disable unnecessary new features or mitigate any additional security exposure.
Vulnerability management type
Supplier-defined controls
Vulnerability management approach
We monitor vulnerability feeds, vendor and security researcher blogs to understand emerging threats. We then make an assessment to confirm any potential to affect service, and if so, determine severity and likelihood. Based on our assessments, we create a mitigation plan that may include a change to operation practice, a planned upgrade or an emergency upgrade. Operational changes and emergency upgrades are carried out as soon as reasonably practical following assessment and planning. Planned upgrades are accommodated within the normal upgrade or patching cycle.
Protective monitoring type
Supplier-defined controls
Protective monitoring approach
System event monitoring is used to facilitate the identification, classification and analysis of potential security incidents. Any security-related incident is classified as a severely service affecting incident, is escalated accordingly and actioned immediately. Different techniques and processes are employed to mitigate and recover service integrity, depending on the nature of the incursion. Further details are available on request.
Incident management type
Supplier-defined controls
Incident management approach
All service incidents follow a standard process. A triage step classifies to: common event not service affecting; service affecting; severely service affecting. Common events are handled by following a routine process. Events affecting service severely are immediately escalated to acquire necessary resource and management support. All other incidents are handled by the respective support team. Clients can report incidents by email to the support team. Update reports and communications are issued for all client-reported and service affecting incidents. When service affecting incidents are resolved, root cause analysis is performed to determine mitigating actions.

Secure development

Approach to secure software development best practice
Supplier-defined process

Separation between users

Virtualisation technology used to keep applications and users sharing the same infrastructure apart
Who implements virtualisation
Virtualisation technologies used
How shared infrastructure is kept separate
While cloud resources are fully virtualised for flexibility and scale, our security is hardware based and implemented at the network layer directly on Cisco switches, allowing both high security and high performance. Individual cloud compute and storage resources are accessed via secure VLAN segmentation on top of this configuration. The combination of VMware vSphere hypervisor and high speed access to block storage via fibre channel allows full customisation of cloud resources. Security and network services available include customisable firewalls, network address translation, virtual IP addresses, load balancing and multicast.

Energy efficiency

Energy-efficient datacentres
Description of energy efficient datacentres
Our data centre provider is a Corporate Participant in the European Code of Conduct for Energy Efficiency in Data Centres programme and recently won an award from the COC for adopting innovative technologies to improve data centre energy efficiency and reduce energy consumption. An example is the use of fuel cells that use a proprietary solid oxide technology to generate electricity through a clean electrochemical process. The process significantly reduces carbon emissions, saves water that would have been used in conventional power generation, and virtually eliminates harmful oxides and particulate emissions. All our data centres have ISO 50001 certification, awarded by Lloyd's Register Quality Assurance, which requires an ongoing commitment to energy management and continuous improvement in energy performance. We are also certified for ISO 14001, which provides assurance that environmental impact is being measured and improved. Our data centres are also ENERGY STAR certified. ENERGY STAR is a U.S. Environmental Protection Agency programme that assesses how efficiently a facility uses energy compared to similar facilities. A calculation similar to power usage effectiveness (PUE) is among the metrics considered for ENERGY STAR data centre certification. Additional specifics are available on request to buyers as part of the procurement process.


£24,600 a terabyte a year
Discount for educational organisations
Free trial available

Service documents