0.01
No commit activity in last 3 years
No release in over 3 years
There's a lot of open issues
Easily cache objects in Ruby, using a Redis store backend
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
 Dependencies

Development

~> 2.1
~> 1.5
~> 0.16
>= 0
>= 0
 Project Readme

Object::Cache wercker status

Easy caching of Ruby objects, using Redis as a backend store.

  • Installation
  • Quick Start
  • Usage
    • marshaling data
    • ttl
    • namespaced keys
    • key prefixes
    • redis replicas
    • core extension
  • License

Installation

Add this line to your application's Gemfile:

gem 'object-cache'

And then execute:

bundle

Or install it yourself as:

gem install object-cache

Quick Start

# require the proper libraries in your project
require 'redis'
require 'object/cache'

# set the backend to a new Redis instance
Cache.backend = Redis.new

# wrap your object in a `Cache.new` block to store the object on first usage,
# and retrieve it again on subsequent usages
Cache.new { 'hello world' }

# add the core extension for easier access
require 'object/cache/core_extension'
cache { 'hello world' }

Usage

Using Object::Cache, you can cache objects in Ruby that have a heavy cost attached to initializing them, and then replay the recorded object on any subsequent requests.

For example, database query results can be cached, or HTTP requests to other services within your infrastructure.

Caching an object is as easy as wrapping that object in a Cache.new block:

Cache.new { 'hello world' }

Here, the object is of type String, but it can be any type of object that can be marshalled using the Ruby Marshal library.

marshaling data

You can only marshal data, not code, so anything that produces code that is executed later to return data (like Procs) cannot be cached. You can still wrap those in a Cache.new block, and the block will return the Proc as expected, but no caching will occur, so there's no point in doing so.

ttl

By default, a cached object has a ttl (time to live) of one week. This means that every request after the first request uses the value from the cached object. After one week, the cached value becomes stale, and the first request after that will again store the (possibly changed) object in the cache store.

You can globaly set the default ttl to a different value:

Cache.default_ttl = 120

You can easily modify the ttl per cached object, using the keyword argument by that same name:

Cache.new(ttl: 60) { 'remember me for 60 seconds!' }

Or, if you want the cached object to never go stale, disable the TTL entirely:

Cache.new(ttl: nil) { 'I am forever in your cache!' }
Cache.new(ttl: 0) { 'me too!' }

Note that it is best to never leave a value in the backend forever. Since this library uses file names and line numbers to store the value, a change in your code might mean a new cache object is created after a deployment, and your old cache object becomes orphaned, and will polute your storage forever.

namespaced keys

When storing the key/value object into Redis, the key name is based on the file name and line number where the cache was initiated. This allows you to cache objects without specifying any namespacing yourself.

If however, you are storing an object that changes based on input, you need to add a unique namespace to the cache, to make sure the correct object is returned from cache:

Cache.new(email) { User.find(email: email) }

In the above case, we use the customer's email to correctly namespace the returned object in the cache store. The provided namespace argument is still merged together with the file name and line number of the cache request, so you can re-use that same email namespace in different locations, without worrying about any naming collisions.

key prefixes

By default, the eventual key ending up in Redis is a 6-character long digest, based on the file name, line number, and optional key passed into the Cache object:

Cache.new { 'hello world' }
Cache.backend.keys # => ["22abcc"]

This makes working with keys quick and easy, without worying about conflicting keys.

However, this does make it more difficult to selectively delete keys from the backend, if you want to purge the cache of specific keys, before their TTL expires.

To support this use-case, you can use the key_prefix attribute:

Cache.new(key_prefix: 'hello') { 'hello world' }
Cache.backend.keys # => ["hello_22abcc"]

This allows you to selectively purge keys from Redis:

keys = Cache.backend.keys('hello_*')
Cache.backend.del(keys)

You can also use the special value :method_name to dynamically set the key prefix based on where the cached object was created:

Cache.new(key_prefix: :method_name) { 'hello world' }
Cache.backend.keys # => ["test_key_prefix_method_name_22abcc"]

Or, use :class_name to group keys in the same class together:

Cache.new(key_prefix: :class_name) { 'hello world' }
Cache.backend.keys # => ["CacheTest_22abcc"]

You can also define these options globally:

Cache.default_key_prefix = :method_name

redis replicas

Before, we used the following setup to connect Object::Cache to a redis backend:

Cache.backend = Redis.new

The Ruby Redis library has primary/replicas support built-in using Redis Sentinel.

If however, you have your own setup, and want the writes and reads to be separated between different Redis instances, you can pass in a hash to the backend config, with a primary and replicas key:

Cache.backend = { primary: Redis.new, replicas: [Redis.new, Redis.new] }

When writing the initial object to the backend, the primary Redis is used. On subsequent requests, a random replica is used to retrieve the stored value.

The above example obviously only works if the replicas receive the written data from the primary instance.

core extension

Finally, if you want, you can use the cache method, for convenient access to the cache object:

require 'object/cache/core_extension'

# these are the same:
cache('hello', ttl: 60) { 'hello world' }
Cache.new('hello', ttl: 60) { 'hello world' }

That's it!

License

The gem is available as open source under the terms of the MIT License.