Frankie Dog Hosting leverages Amazon Web Services Global Infrastructure to provide unmatched speeds, security, availability and scalability. Using our proprietary technology specifically developed for WordPress hosting, we offer users the newest and best hosting experience.
Frankie Fact #1
The Best Infrastructure.
We use Amazon EC2 servers. Amazon EC2 servers offer the largest global network with the fastest processors in the cloud.
Frankie Fact #2
Global Accelerator = Fast.
Amazon Global Accelerator sends your traffic through Amazon web service’s global network infrastructure, improving your internet user performance by up to 60%.
Frankie Fact #3
ELB = Speed + Security.
Amazon Elastic Load Balancing ( Amazon ELB) automatically distributes incoming application traffic across multiple targets, providing high availability, automatic scaling and robust security.
Frankie Fact #4
EKS = Scale + Security.
Amazon Amazon Elastic Kubernetes Service (Amazon EKS) helps us provide highly-available and secure clusters for your WordPress site. Customers such as Intel, Snap, Intuit, GoDaddy, and Autodesk trust EKS to run their most sensitive and mission critical applications.
Amazon EC2 Servers (EC2)
Amazon Elastic Compute Cloud (Amazon EC2) is a web service that provides secure, resizable compute capacity in the cloud. It is designed to make web-scale cloud computing easier for developers. Amazon EC2’s simple web service interface allows you to obtain and configure capacity with minimal friction. It provides you with complete control of your computing resources and lets you run on Amazon’s proven computing environment.
Amazon EC2 offers the broadest and deepest compute platform with choice of processor, storage, networking, operating system, and purchase model. We offer the fastest processors in the cloud and we are the only cloud with 400 Gbps ethernet networking. We have the most powerful GPU instances for machine learning training and graphics workloads, as well as the lowest cost-per-inference instances in the cloud. More SAP, HPC, Machine Learning, and Windows workloads run on AWS than any other cloud.
- Reliable, scalable, infrastructure on demand
- Increase or decrease capacity within minutes, not hours or days
- SLA commitment of 99.99% availability for each Amazon EC2 region.
- Each region consists of at least 3 availability zones
- The AWS Region/AZ model has been recognized by Gartner as the recommended approach for running enterprise applications that require high availability
Amazon Global Accelerator
AWS Global Accelerator is a networking service that sends your user’s traffic through Amazon Web Service’s global network infrastructure, improving your internet user performance by up to 60%.
Accelerate latency sensitive applications: AWS Global Accelerator combines advanced networking features with the dedicated AWS Global Network to improve your application network performance by up to 60%.
Improve resiliency and availability: Wherever you route your traffic on the AWS network, with Global Accelerator, failover between application origins happens automatically and within seconds. If Global Accelerator detects failure of your application endpoint it instantly triggers traffic re-routing to the next available, closest endpoint in another AZ or AWS Region.
Simplified global traffic management: AWS Global Accelerator simplifies global traffic management by providing 2 static anycast IP addresses that only need to be allow-listed by users once. Behind these IP address you can add or remove AWS origins, opening up uses such as endpoint failover, scaling, or testing without any user-side changes.
Protect your applications: Exposing your AWS origins, such as Application Load Balancers or EC2 instances, to public internet traffic creates an opportunity for malicious attacks. AWS Global accelerator decreases the risk of attack by masking your origin behind two static entry points. These entry points are protected by default from Distributed Denial of Service (DDoS) attacks with AWS Shield.
Amazon Elastic Load Balancing (ELB)
Elastic Load Balancing automatically distributes incoming application traffic across multiple targets, such as Amazon EC2 instances, containers, IP addresses, Lambda functions, and virtual appliances. It can handle the varying load of your application traffic in a single Availability Zone or across multiple Availability Zones. Elastic Load Balancing offers four types of load balancers that all feature the high availability, automatic scaling, and robust security necessary to make your applications fault tolerant.
Highly availability and elasticity: Elastic Load Balancing is part of the AWS network, with native awareness of failure boundaries like AZs to keep your applications available across a region, without requiring Global Server Load Balancing (GSLB). ELB is also a fully managed service, meaning you can focus on delivering applications and not installing fleets of load balancers. Capacity is automatically added and removed based on the utilization of the underlying application servers.
Security: Elastic Load Balancing works with Amazon Virtual Private Cloud (VPC) to provide robust security features, including integrated certificate management, user-authentication, and SSL/TLS decryption. Together, they give you the flexibility to centrally manage TLS settings and offload CPU intensive workloads from your applications. ALB also supports integration with AWS WAF, adding a level of protection before bad actors reach the application. Further, S2N and HTTP Guardian have been developed as Open Source solutions to reduce the potential for HTTP-based attacks.
Feature breadth: Elastic Load Balancing offers the breadth of features needed by businesses of all sizes, while delivering them in an AWS-native experience. Elastic Load Balancing includes support for features needed in container-based workloads, including HTTP/2, gRPC, TLS offload, advanced rule-based routing, and integration with container services as an ingress controller. ALB provides customers with a native HTTP endpoint for calling Lambda functions, removing the dependency on other solutions. Further, Gateway Load Balancer creates one gateway for routing traffic through fleets of third-party appliances.
Robust monitoring & visibility: Elastic Load Balancing allows you to monitor the health of your applications and their performance in real time with Amazon CloudWatch metrics, logging, and request tracing. This improves visibility into the behavior of your applications, uncovering issues and identifying performance bottlenecks in your application stack. ELB helps ensure compliance with application Service Level Agreements (SLAs).
Integration and global reach: As a native AWS service, ELB is tightly integrated with other AWS services like EC2, ECS/EKS, Global Accelerator and operational tools such as AWS CloudFormation and AWS Billing. Across the Amazon Global Infrastructure and customer data centers with AWS Outposts and on-premises target support, ELB is available everywhere you run your AWS workloads.
Amazon Amazon Elastic Kubernetes Service (EKS)
Amazon Elastic Kubernetes Service (Amazon EKS) gives you the flexibility to start, run, and scale Kubernetes applications in the AWS cloud or on-premises. Amazon EKS helps you provide highly-available and secure clusters and automates key tasks such as patching, node provisioning, and updates. Customers such as Intel, Snap, Intuit, GoDaddy, and Autodesk trust EKS to run their most sensitive and mission critical applications.
Improve availability and observability: EKS runs the Kubernetes control plane across multiple AWS Availability Zones, automatically detects and replaces unhealthy control plane nodes, and provides on-demand, zero downtime upgrades and patching. EKS offers a 99.95% uptime SLA. At the same time, the EKS console provides observability of your Kubernetes clusters so you can identify and resolve issues faster.
Provision and scale your resources efficiently: With EKS managed node groups, you don’t need to separately provision compute capacity to scale your Kubernetes applications. You can also choose AWS Fargate to automatically provision on-demand serverless compute for your applications. For even more cost savings, EKS nodes on Amazon EC2 Spot instances reduce your cost for more efficiency.
Get a more secure Kubernetes environment: EKS automatically applies the latest security patches to your cluster’s control plane. AWS works closely with the community to address critical security issues and help ensure that every EKS cluster is secure.
Amazon ElastiCache for Redis is a blazing fast in-memory data store that provides sub-millisecond latency to power internet-scale real-time applications. Built on open-source Redis and compatible with the Redis APIs, ElastiCache for Redis works with your Redis clients and uses the open Redis data format to store your data. Your self-managed Redis applications can work seamlessly with ElastiCache for Redis without any code changes. ElastiCache for Redis combines the speed, simplicity, and versatility of open-source Redis with manageability, security, and scalability from Amazon to power the most demanding real-time applications in Gaming, Ad-Tech, E-Commerce, Healthcare, Financial Services, and IoT.
Extreme performance: Amazon ElastiCache for Redis works as an in-memory data store to support the most demanding applications requiring sub-millisecond response times. In comparison to disk-based databases where most operations need a roundtrip to disk, in-memory data stores manage data in-memory which is an order of magnitude faster than disks. The result is blazing fast performance with average read or write operations taking less than a millisecond and support for hundreds of millions of operations per second within a cluster. ElastiCache gives you an optimized end-to-end hardware and software stack for blazing fast performance.
Secure: Starting with Amazon ElastiCache for Redis 6, ElastiCache now provides you with the ability to create and manage users and user groups that can be used to set up Role-Based Access Control (RBAC) for Redis commands. You can now simplify your architecture while maintaining security boundaries and also take advantage of granular access control to manage groups. Amazon ElastiCache for Redis supports Amazon VPC, so you can isolate your cluster to the IP ranges you choose for your nodes. The ElastiCache team continuously monitors for known security vulnerabilities in open-source Redis, operating system, and firmware to keep your Redis environment secure. It is PCI compliant, HIPAA eligible, FedRAMP authorized, and offers encryption in transit, at rest (including customer managed CMK stored in AWS KMS), and Redis AUTH for secure internode communications to help keep sensitive data such as personally identifiable information (PII) safe.
Fully managed & hardened: Amazon ElastiCache for Redis is a fully managed service. You no longer need to perform management tasks such as hardware provisioning, software patching, setup, configuration, monitoring, failure recovery, and backups. ElastiCache continuously monitors your clusters to keep your Redis up and running so that you can focus on higher value application development. It provides detailed monitoring metrics of your Redis usage, enabling you to track application trends and adjust cluster configuration, as needed. ElastiCache adds automatic write throttling, intelligent swap memory management, and failover enhancements to improve upon the availability and manageability of open source Redis.
Redis-compatible: Redis is a widely adopted in-memory data store for use as a database, cache, message broker, queue, session store, and leaderboard. Amazon ElastiCache for Redis maintains compatibility with the open source Redis data formats, Redis APIs, and works with Redis clients. You can migrate your self-managed Redis workloads to ElastiCache for Redis without any code change.
Highly available and reliable: Amazon ElastiCache for Redis supports both Redis cluster and non-cluster modes and provides high availability via support for automatic failover by detecting primary node failures and promoting a replica to be primary with minimal impact. It allows for read availability for your application by supporting read replicas (across availability zones), to enable the reads to be served when the primary is busy with the increased workload. ElastiCache for Redis supports enhanced failover logic to allow for automatic failover in cases when majority of the primary nodes for Redis cluster mode are unavailable. On Redis 5.0.5 onwards, auto-failover enabled clusters provide online configuration changes for all planned operations.
Easily scalable: With Amazon ElastiCache for Redis, you can start small and easily scale your Redis data as your application grows – all the way up to a cluster with 340 TB of in-memory data. It allows you to scale your Redis Cluster environment up to 500 nodes and 500 shards. It supports online cluster resizing to scale-out and scale-in your Redis clusters without downtime and adapts to changing demand. To scale read capacity, ElastiCache allows you to add up to five read replicas across multiple availability zones. To scale write capacity, ElastiCache supports Redis cluster mode which enables you to partition your write traffic across multiple primaries.