Go back to fronty page View most popular entries View latest additions Submit tutorials to UnixTutorials.info
UnixTutorials logo

Search results for Configuring Apache for Maximum Performance

Debian

Setting Up A High-Availability Load Balancer (With Failover and Session Support) With HAProxy/Heartbeat On Debian Etch

Post date: November 7, 2007, 11:11 Category: Network Views: 5119 Comments
Tutorial quote: This article explains how to set up a two-node load balancer in an active/passive configuration with HAProxy and heartbeat on Debian Etch. The load balancer sits between the user and two (or more) backend Apache web servers that hold the same content. Not only does the load balancer distribute the requests to the two backend Apache servers, it also checks the health of the backend servers. If one of them is down, all requests will automatically be redirected to the remaining backend server. In addition to that, the two load balancer nodes monitor each other using heartbeat, and if the master fails, the slave becomes the master, which means the users will not notice any disruption of the service. HAProxy is session-aware, which means you can use it with any web application that makes use of sessions (such as forums, shopping carts, etc.).
Debian

Setting Up A High-Availability Load Balancer (With Failover and Session Support) With HAProxy/Keepalived On Debian Etch

Post date: October 29, 2007, 09:10 Category: Network Views: 5265 Comments
Tutorial quote: This article explains how to set up a two-node load balancer in an active/passive configuration with HAProxy and keepalived on Debian Etch. The load balancer sits between the user and two (or more) backend Apache web servers that hold the same content. Not only does the load balancer distribute the requests to the two backend Apache servers, it also checks the health of the backend servers. If one of them is down, all requests will automatically be redirected to the remaining backend server. In addition to that, the two load balancer nodes monitor each other using keepalived, and if the master fails, the slave becomes the master, which means the users will not notice any disruption of the service. HAProxy is session-aware, which means you can use it with any web application that makes use of sessions (such as forums, shopping carts, etc.).
Linux

Building a Linux Cluster, Part 1

Post date: April 18, 2005, 03:04 Category: Network Views: 2873 Comments
Tutorial quote: Linux clusters have escaped. As the general popularity of the Linux operating system increases, more complex solutions built with it are becoming increasingly common in the "traditional" (more conservative) IT world. Linux computer clusters, whose provenance was originally universities and research institutions such as the U.S. National Laboratories, are showing up in increasing numbers as high-performance computing solutions within such areas as oil and gas exploration, computer-aided engineering, visualization, and software development. Linux clusters providing highly-available web, mail, and other infrastructure services are also increasingly common. If past computing history is any indicator of future trends, widespread use of Linux clusters in the mainstream IT world cannot be far behind.
Debian

Simple Apache 2 Tomcat 5 mod_jk Integration

Post date: April 13, 2006, 07:04 Category: Network Views: 3772 Comments
Tutorial quote: The whole tutorial is based on many tutorials, but I made a very simple one, with no virtual hosts.
Linux

Setting Up Subversion for One or Multiple Projects

Post date: April 13, 2005, 00:04 Category: Miscellaneous Views: 4530 Comments
Tutorial quote: Deploying a secure and manageable Subversion installation that uses Apache 2.0 as a central authentication checkpoint and SSL for data integrity and confidentiality.
Unix+clones

Installing and securing Squid

Post date: March 13, 2006, 15:03 Category: Software Views: 5464 Comments
Tutorial quote: Squid is a high-performance proxy caching server for web clients, supporting FTP, gopher, and HTTP data objects. Unlike traditional caching software, Squid handles all requests in a single, non-blocking, I/O-driven process. Squid keeps meta data and especially hot objects cached in RAM, caches DNS lookups, supports non-blocking DNS lookups, and implements negative caching of failed requests. Squid supports SSL, extensive access controls, and full request logging. By using the lightweight Internet Cache Protocol, Squid caches can be arranged in a hierarchy or mesh for additional bandwidth savings.

After the installation and base configuration of squid we will add another layer of security by chrooting it.
Linux

Installing and Using Hadoop

Post date: November 28, 2009, 22:11 Category: System Views: 4980 Comments
Tutorial quote: Learn to install configure and use the open source Hadoop framework from the Apache Foundation. Basic installation, configuration, executing jobs, and using the HDFS filesystem and the Web GUI.
Debian

Creating self signed SSL Certificates Using Openssl

Post date: February 3, 2006, 01:02 Category: Network Views: 3133 Comments
Tutorial quote: Easy howto on how to create self signed ssl certificate for apache.
RedHat

Choosing an I/O Scheduler for Red Hat Enterprise Linux 4

Post date: July 18, 2005, 22:07 Category: Benchmarks Views: 8356 Comments
Tutorial quote: The Linux kernel, the core of the operating system, is responsible for controlling disk access by using kernel I/O scheduling. Red Hat Enterprise Linux 3 with a 2.4 kernel base uses a single, robust, general purpose I/O elevator. The 2.4 I/O scheduler has a reasonable number of tuning options by controlling the amount of time a request remains in an I/O queue before being serviced using the elvtune command. While Red Hat Enterprise Linux 3 offers most workloads excellent performance, it does not always provide the best I/O characteristics for the wide range of applications in use by Linux users these days. The I/O schedulers provided in Red Hat Enterprise Linux 4, embedded in the 2.6 kernel, have advanced the I/O capabilities of Linux significantly. With Red Hat Enterprise Linux 4, applications can now optimize the kernel I/O at boot time, by selecting one of four different I/O schedulers.
Debian

Setting up an SSL server with Apache2

Post date: February 12, 2006, 07:02 Category: Network Views: 4918 Comments
Tutorial quote: With the introduction of the Apache2 packages in Debian it is much simpler to create and use a secure SSL protected webserver than in the old days with Apache 1.3, here we'll show how it is done.
Web-based applications and online marketing solutions - LumoLink