Distributed search for a counterexample to Beal's Conjecture!

Beal's Conjecture says that if a^x + b^y = c^z, where a, b, c, x, y, and z are positive integers and x, y and z are all greater than 2, then a, b, and c must have a common prime factor.

There is a monetary prize offered by Andrew Beal for a proof or counterexample to the conjecture. More information about the prize can be found here http://www.ams.org/profession/prizes-awards/ams-supported/beal-prize. This project aims to expand the size of the counterexample search space covered compared to previous efforts by using a distributed search strategy.

Previous Efforts

Current Status

Where are we at with the search? Below is a matrix describing the state of known past attempts, as well as my own using the distributed approach described here. Below N represents the spaces covered by Peter Norvig (http://norvig.com/beal.html), and the D at 1000x1000 was covered at http://www.danvk.org/wp/beals-conjecture/. The cells marked CURRENT are the jobs I've performed, and the results are available in the results/ directory of this Git repository.

p=7 p=10 p=30 p=100 p=1,000 p=2000 p=3,000
b=100 N N N N N
b=1,000 N N N N D
b=2,000 -- -- -- -- -- CURRENT
b=3,000 N N N N -- -- CURRENT
b=10,000 N N N N
b=100,000 N N -- CURRENT
b=250,000 N -- -- CURRENT

Problem Overview

The conceptual strategy for conducting a counterexample search is to compute all possible points (a, x, b, y, c, z) and then evaluate the expression a^x + b^y = c^z. If the expression holds and the bases do not have a common prime factor, then a counterexample has been found.

The core challenge behind a counterexample search strategy is dealing with the enormous size of the space being examined. For instance, taking a maximum value for the bases and exponents of 1000 we end up with 1000^6 points. That number is so large that a 1GHz CPU that runs at 1 billion cycles per second will take 1 billion seconds just to execute 1000^6 1 cycle instructions. Evaluating the expression a^x + b^y = c^z takes far more than 1 cycle, so we need to be smart about the search.

First we will describe the simplest algorithm for conducting a counterexample search, and then iteratively refine it with various optimizations. Then we will show how scaling this simple algorithm hits a wall very quickly, and then describe additional optimizations that let us expand the search space past what is possible with the simple algorithm. The optimizations and search techniques that will be described have all been considered in previous efforts. Finally we'll describe our new implementation that distributes the problem allowing us to further scale the problem sizes being considered.

Generation 1 Algorithm

The following is a naive algorithm for conducting a search that computes every combination of the points (a, x, b, y, c, z) and then checks if the point represents a counterexample:

for a in range(1, max_base+1):
    for b in range(1, max_base+1):
        for x in range(3, max_pow+1):
            for y in range(3, max_pow+1):
                for c in range(1, max_base+1):
                    for z in range(3, max_pow+1):
                        check(a, x, b, y, c, z)

A candidate version of the check function might do something like the following:

def check(a, x, b, y, c, z):
    axby, cz = pow(a, x) + pow(b, y), pow(c, z)
    if axby == cz and gcd(a, b) == 1 and gcd(a, c) == 1 and gcd(b, c) == 1:
        print "found counterexample:", a, x, b, y, c, z

I ran this algorithm for 85 minutes. In that period of time 859,796,767 points were examined. Considering the size of the state space with maximum base and exponent values of 1000, that is approximately 0.0000000859796767 % of the total space covered. In reality it will probably run much slower because this short experiment didn't reach the very large exponents that take a lot of effort to compute. This test was written in Python, but even when written in optimized C, this strategy will simply not scale with search spaces this size. In order to make progress we need to cut down on the amount of work we are doing, as well as making the operations more efficient.

Optimization 1

Since a^x + b^y is commutative we don't have to bother also testing b^y + a^x. This can be incorporated into the algorithm above by adjusting the upper bound of the values assigned to b:

for a in range(1, max_base+1):
    for b in range(1, a+1):

This filter reduces by 50% the number of points that must be considered. The problem is that the search space is so large that even with a 50% reduction in size it is still much too large. But we certainly want to use any optimizations we can find.

Optimization 2

Since we only care about counterexamples we don't need to check any points where the bases have a common prime. This means that when iterating over the points in the space we can skip all points for which any two of the bases (e.g. a and b) have a common prime factor.

for a in range(1, max_base+1):
    for b in range(1, a+1):
        if fractions.gcd(a, b) > 1:

This is actually a pretty nice optimization. Even though computing gcd isn't cheap, we only have to do it roughly max_base * (max_base + 1) / 2 times, and what we get in return is the ability to completely skip the rest of the nested for loops for that combination of (a, b) values.

Optimization 3

Notice that in the algorithm above that computes all combinations of (a, x, b, y, c, z) the outer four for loops compute the possible values of a^x + b^y (the left-hand-side), and the inner most two for loops compute all possible values of c^z to which a particular left-hand-side value is compared. Recomputing all possible c^z values is very expensive! An alternative is to pre-compute all possible values for c^z and save them in a data structure that can be searched. Then, given a left-hand-side value we can search for the corresponding match among the pre-computed values. If that search is cheaper than re-computing the values then we will speed-up the overall search.

The following is a revised version of our algorithm that pre-computes the c^z values and stores them in a Python dictionary. Note that below the main search loop only generates combinations of a^x + b^y, significantly reducing the amount of work we have to do.

cz_values = defaultdict(lambda: [])

# populate c^z table: one time cost
for c in range(1, max_base+1):
    for z in range(3, max_pow+1):
        value = pow(c, z)
        cz_values[value].append((c, z))

# generate left-hand-side values
for a in range(1, max_base+1):
    for b in range(1, a+1):
        if gcd(a, b) > 1:
        for x in range(3, max_pow+1):
            for y in range(3, max_pow+1):
                check(a, x, b, y)

And now the check function can avoid computing all c^z values, opting instead for a fast hash-table lookup:

def check(a, x, b, y):
    axby = pow(a, x) + pow(b, y)
    if axby in cz_values: # avoid empty list creation
        for c, z in cz_values[axby]:
            if gcd(a, b) == 1 and gcd(a, c) == 1 and gcd(b, c) == 1:
                print "counterexample found:", a, x, b, y, c, z

It may cost a few seconds to pre-compute all of the c^z values for a very large search space, but since we get to amortize that cost across the all points in the space we effectively reduce the cost of the search by a factor proportional to the size of the c^z space!

Optimization 4

Notice that to compute the values of c^z to store in a search structure we also compute all of the powers that are needed for evaluating a^x + b^y. Since pow isn't a cheap function, we can also save each value to avoid recomputing terms in the equation. Here we modify the way we populate the c^z search structure to also cache the individual powers:

cz_values = defaultdict(lambda: []) # pow(c, z) -> { (c, z) }
powers = defaultdict(lambda: {})    #    (c, z) -> pow(c, z)

for c in range(1, max_base+1):
    for z in range(3, max_pow+1):
        value = pow(c, z)
        cz_values[value].append((c, z))
        powers[c][z] = value

Then in the check function we lookup each a^x and b^y to avoid the expensive computation of the pow function.

def check(a, x, b, y):
    axby = powers[a][x] + powers[b][y]
    if axby in cz_values: # avoid empty list creation
        for c, z in cz_values[axby]:
            if gcd(a, b) == 1 and gcd(a, c) == 1 and gcd(b, c) == 1:
                print "counterexample found:", a, x, b, y, c, z

Now we are getting somewhere. This approach closely resembles the approach used by Peter Norvig here http://norvig.com/beal.html, with the exception that there are some more efficient ways to cache data and build the search structure. Even in Python, this approach is pretty speedy. Norvig reports results for several different search spaces, including 3 minutes for max_base = max_pow = 100. Unfortunately the costs explode with large spaces such as 19 hours for max_pow = 1000 and max_base = 100, and 39 days max_pow = 100 and max_base = 10,000.

We can distribute the search problem by assigning worker nodes disjoint partitions of the search space. However, the mechanism we have described uses infinite precision arithmetic which adds a lot of overhead both in time and space, limiting per-node scalability. For instance, the value 1000^1000 contains about 3000 digits. Operations on numbers this large can't be performed as efficiently compared to numbers that are stored in 64-bit registers.

In the previous approach described here http://norvig.com/beal.html, Peter Norvig proposed doing all arithmetic modulo large 64-bit prime numbers. This has the advantage that all operations are very efficient, but results may be false positives and must be verified. However, if we can sufficiently reduce the number of potential counterexamples, the savings realized from performing verification using infinite precision arithmetic on a small percentage of the total space may be far outweighed by the cost of using infinite precision arithmetic exclusively. This is exactly the approach taken in http://www.danvk.org/wp/beals-conjecture/. Next I'll describe how to incorporate modulo arithmetic to make the search more efficient.

Generation 2 Algorithm

In the previous section we described the 1st generation algorithm and a set of optimizations that significantly improve performance over a naive implementation. However, the 1st generation algorithm used infinite precision arithmetic which limits scalability because of the costs associated with operations on large numbers. Performing modulo arithmetic can circumvent the problem.

The motivation for using modulo arithmetic is to reduce the number of bits needed to represent the values we are working with. While space savings is important, numeric operations on values that fit into CPU registers are extremely fast compared to the algorithms used to perform infinite precision arithmetic. The approach is based off the observation that if a^x + b^y = c^z then for a value p1, (a^x + b^y) mod p1 = c^z mod p1. By keeping p1 small enough, we can force all values to fit within a CPU register. For instance, 789^999 is a number with 2895 digits. However, 789^999 mod 4294967291 is 1153050910 which easily fits into a 64-bit CPU register.

The obvious problem with sacrificing precision is false positives. That is, the space of possible values is far greater than the size of the modulo meaning that many distinct values may be identical modulo the same number. So how do we reduce the number of false positives? Observing that if two numbers are identical modulo a value p1 then they are identical modulo a different value p2, we can construct a simple filter by performing arithmetic modulo multiple numbers and ensuring that equality holds in all cases. There is a more detailed description of this approach written about here http://www.danvk.org/wp/beals-conjecture/, which also describes why large prime numbers are good choice for the modulus.

The changes to the algorithm are relatively modest. Each of the search structures used to store and cache the c^z values is parameterized on the prime numbers. So for instance to look up 10^20 mod 4294967291 we use powers[4294967291][10][20].

primes = [4294967291, 4294967279]
cz_values = {}
powers = {}

# parameterize structures on prime values
for prime in primes:
    cz_values[prime] = defaultdict(lambda: [])
    powers[prime] = defaultdict(lambda: {})

# compute c^z modulo each prime value
for c in range(1, max_base+1):
    for z in range(3, max_pow+1):
        for prime in primes:
            value = pow(c, z, prime)
            cz_values[prime][value].append((c, z))
            powers[prime][c][z] = value

The loop construct used to generate all of the (a, x, b, y) values is the same, and the remaining changes are in the check function where we ensure that equality holds modulo each prime number.

def check(a, x, b, y):
    candidates = []
    candidates_found = True
    for prime in primes:
        axby = (powers[prime][a][x] + powers[prime][b][y]) % prime
        if not axby in cz_values[prime]:
            candidates_found = False # done... needs to be true for all primes
        for c, z in cz_values[prime][axby]:
            candidates.append((a, x, b, y, c, z))
    if candidates_found:
        print candidates

This is quite a lot faster than the generation 1 algorithm. We can search the space associated the parameters max_base = max_pow = 100 in about 20 seconds, compared to 3 minutes with the previous version. The time savings is realized by using small fixed-size numbers.

In the next section we describe an implementation in C++ that is designed to be distributed such that disjoint partitions of the search space are handled by independent workers, allowing us to scale out the search and reduce the costs of searching larger spaces.

Generation 3

The implementation found in this repository contains all of the above optimizations, but adds the ability to decompose the problem and perform the search in parallel across any number of worker nodes. The high-level structure of the solution is for each worker node to host a copy of the c^z space such that it can examine any partitioning of the a^x + b^y space. Since the c^z space is relatively small, we try much larger spaces without worrying about memory pressure. A master node computes partitions, responds to requests for work, and records results from worker nodes.


The manager process (prob-manager.py) exposes two RPC endpoints get_work and finish_work. The partitioning of the a^x + b^y space is very simple, handing out distinct values for a (smarter partitioning is needed for expanding the search in certain ways, but this works for now). When a worker requests a partitioning the the master process will run the following, where __get_work() returns a distinct a value:

def get_work(self):
    with self._lock:
        part = self.__get_work()
        if not part:
            return None
        work_spec = {'max_base': self._maxb,
                'max_pow': self._maxp,
                'primes': self._primes,
                'part': part}
        return work_spec

When a worker completes a search it responds to the finish_work RPC endpoint. A completed work unit contains enough information to detect duplicates, as well as all of the candidate solutions that the worker found. The master writes all of the candidates out to a single file.

def finish_work(self, spec, results):
    with self._lock:
        dupe = self._work_queue.complete(spec)
        if dupe:
        if self._output:
            with open(self._output, 'a') as f:
                for result in results:
                    print result
                    f.write("%d %d %d %d\n" % tuple(result))
            print (spec, results)


The worker process (prob-worker.py) is responsible for handling a partition of the a^x + b^y space. The following code snippet is run by each worker. In an infinite loop a work unit is retrieved, a context is created (see below), a search is performed, and the results are returned to the manager.

def run(self):
    while True:
        work_spec = self._server.get_work()
        if not work_spec:
            print "no work available... waiting"
        part = work_spec['part']
        hits = search.search(part[0])
        self._server.finish_work(part, tuple(hits))

Note above that for each work unit a new context is created (i.e. setup_context()). A context contains all of the data structures used to perform a search and can take some time to setup. Currently a worker guarantees that a context is compatible with each work unit, but doesn't attempt to switch contexts after the first work unit is retrieved. This is primarily just for simplicity:

search = None

def setup_context(work_spec):
    global search
    max_base = work_spec['max_base']
    max_pow = work_spec['max_pow']
    primes = work_spec['primes']
    if not search:
        search = beal.search(max_base, max_pow, primes)
    assert max_base == search.max_base()
    assert max_pow  == search.max_pow()
    assert primes   == search.primes()

So where is the search actually performed? Well Python can be quite slow at computation, but it is fantastic at tasks such as coordinating work and handling network communication. So we've chosen to implement all the performance critical parts in C.

GCD and Modular Exponentiation

Implementations of GCD and fast modular exponentation are found in math.h. The following were taken from Wikipedia and ported to C from psuedo code. I've only included modpow here:

 * http://en.wikipedia.org/wiki/Modular_exponentiation
static inline uint32_t modpow(uint64_t base, uint64_t exponent, uint32_t mod)
  uint64_t result = 1;

  base = base % mod;
  while (exponent > 0) {
    if (exponent % 2 == 1)
      result = (result * base) % mod;
    exponent = exponent >> 1;
    base = (base * base) % mod;

  assert(result <= (((1ULL) << 32)-1));
  return result;


So how do we go about testing the implementation? After all, these are crucial to the correctness of the search. What we've done is used Python to coordinate a large number of tests and compare the results to a separate implementation of each algorithm.

In beal.cc we have all of the C implementations, and then we bind from Python. First we create a simple C wrapper:

uint32_t c_modpow(uint64_t base, uint64_t exponent, uint32_t mod) {
  return modpow(base, exponent, mod);

unsigned int c_gcd(unsigned int u, unsigned int v) {
  return gcd(u, v);

And in beal.py we create the bindings:

import cffi

_ffi = cffi.FFI()
_libbeal = _ffi.dlopen("./libbeal.so")
uint32_t c_modpow(uint64_t base, uint64_t exponent, uint32_t mod);
unsigned int c_gcd(unsigned int u, unsigned int v);

def modpow(b, e, m):
    return _libbeal.c_modpow(b, e, m)

def gcd(u, v):
    return _libbeal.c_gcd(u, v)

To actually perform the tests we can compute values in Python and compare the values to what is computed by the C versions. In the following snippet __check will use Python built-in function pow and compare to our implementation. We've removed some of the tests here (see test.py), but test_random effectively generates a large number of random values to check. Note the test_specific function where we had actually found a problem in one version of the algorithm on Wikipedia (the fix had already been added after I had snatched the code a while back).

class TestModPow(unittest.TestCase):
    def __check(self, b, e, m):
        value1 = pow(b, e, m)
        value2 = beal.modpow(b, e, m)
        self.assertEqual(value1, value2,
                "%d != %d: %d %d %d" % (value1, value2, b, e, m))

    def test_random(self):
        limit = 10**7
        for _ in range(limit):
            base = random.randint(1, 2**64-1)
            expo = random.randint(1, 2**64-1)
            mod = random.randint(1, 2**32-1)
            self.__check(base, expo, mod)

    def test_specific(self):
        # random testing found a problem with c_modpow for these inputs. the
        # fix was to include `base = base % mod` before starting the loop in
        # the modpow algo, as is done in the wikipedia algorithm.
        self.__check(4542062976100348463, 4637193517411546665, 3773338459)
        self.__check(70487458014159955, 5566498974156504764, 3541295600)

Similarily, to test GCD we use the Python library function fractions.gcd to compute a value to compare to our C version:

class TestGCD(unittest.TestCase):
    def __check(self, u, v):
        value1 = fractions.gcd(u, v)
        value2 = beal.gcd(u, v)
        self.assertEqual(value1, value2, "%u %u" % (u, v))

    def test_dense(self):
        limit = 10**4
        for u in xrange(1, limit):
            for v in xrange(1, limit):
                self.__check(u, v)

The c^z Context

For each prime value used as a filter we create a cz class that contains search structures used to quickly find candidate solutions. The private variable std::vector<std::vector<uint32_t> > vals_ is used to avoid recomputing the exponential terms in a^x + b^y. When performing the actual search we avoid using any type of hash table, and instead exploit the 32-bit bound on prime values by created a 4GB sparse array, represented by the private variable std::vector<bool> exists_;. Note in the constructor that we only set the bits to true that correspond to values found in the c^z space.

Finally, the get and exists methods are used to interface to the data structures.

class cz {
  cz(unsigned int maxb, unsigned int maxp, uint32_t mod) {
    assert(maxb > 0);
    assert(maxp > 2);
    assert(mod > 0);
    for (unsigned int c = 1; c <= maxb; c++) {
      for (unsigned int z = 3; z <= maxp; z++) {
        uint32_t val = modpow(c, z, mod);
        vals_[c][z] = val;
        exists_[val] = true;
    mod_ = mod;

  inline uint32_t get(int c, int z) const {
    assert(c > 0);
    assert(z > 2);
    return vals_[c][z];

  inline bool exists(uint32_t val) const {
    return exists_[val];

  inline uint32_t mod() const {
    return mod_;

  std::vector<std::vector<uint32_t> > vals_;
  std::vector<bool> exists_;
  uint32_t mod_;

To test the cz class we again coordinate things with Python. At a high-level we construct a cz class and then assert various properties. The __check method is responsible for this. First we create the cz instance:

class TestCz(unittest.TestCase):
    def __check(self, maxb, maxp, mod):
        cz = beal.cz(maxb, maxp, mod)

Next we recompute all of the values in Python and ensure that we are able to retrieve the same value through the get interface, and that their value is found with the exists interface:

        values = set()
        for c in xrange(1, maxb+1):
            for z in xrange(3, maxp+1):
                value1 = pow(c, z, mod)
                value2 = cz.get(c, z)
                self.assertEqual(value1, value2)
        assert len(values) > 0

        for value in values:
            exists = cz.exists(value)

Ideally we would like to guarantee that the interface doesn't return false negatives, but this requires querying all values from 0 to 2^32. In the interest of time we compromise by checking a set of random values:

        limit = 10**8
        for _ in xrange(limit):
            value = random.randint(0, 2**32-1)
            exists = cz.exists(value)
            if exists:
                self.assertTrue(value in values)


And that is it. In test.py you can see we call __check with several different sets of parameters. For instance we generate some random parameters:

    def test_random(self):
        limit = 40
        for _ in xrange(limit):
            maxb = random.randint(1, 2000)
            maxp = random.randint(3, 2000)
            mod = random.randint(1, 2**32-1)
            self.__check(maxb, maxp, mod)

Work Context

The context that we setup for a search is primarily just one cz class instance per prime number used as a filter. We organize these instances in the work class. See below in the constructor that we create a cz instance for each prime:

class work {
  work(int maxb, int maxp, uint32_t *primes, size_t nprimes) :
    maxb_(maxb), maxp_(maxp) {
      for (size_t i = 0; i < nprimes; i++) {
        czs_.push_back(new cz(maxb, maxp, primes[i]));

  ~work() {
    for (size_t i = 0; i < czs_.size(); i++)
      delete czs_[i];

The real work happens in do_work that takes as input the partition of a^x + b^y to search (defined currently by a value for a), and returns a set of (a,x,b,y) candidate points as the result. First we create an axby class instance that manages the iteration over the points we will be searching:

  void do_work(int a, std::vector<axby::point>& results) {
    bool done;
    axby pts(maxb_, maxp_, a);

    axby::point& pt = pts.next(&done);
    while (!done) {

Now, for each point and for each prime filter we calculate a^x + b^y mod prime and search for candidates in the corresponding cz instance, collecting the results in an output vector:

      bool found = true;
      for (unsigned i = 0; i < czs_.size(); i++) {
        cz *czp = czs_[i];
        uint64_t ax = czp->get(pt.a, pt.x);
        uint64_t by = czp->get(pt.b, pt.y);
        uint64_t val = (ax + by) % czp->mod();
        if (!czp->exists(val)) {
          found = false;

      if (found)

      pt = pts.next(&done);

  int maxb_;
  int maxp_;
  std::vector<cz*> czs_;

That is a summary of the core approach to the search strategy implementation. The only remaining component is the iteration of the (a,x,b,y) points.

a^x + b^y Point Iteration

We use an axby helper class to iterate over point to check. It is partitioned by values of a. That is, the starting point is a value for the a dimension. All points will be generated and the normal trimming optimizations will be applied. First initialize the class with the starting a value:

class axby {
  struct point {
    point(int a, int x, int b, int y) :
      a(a), x(x), b(b), y(y)

    int a, x, b, y;

  axby(int maxb, int maxp, int a) :
    maxb_(maxb), maxp_(maxp), p_(a, 3, 1, 3), a_dim_(a)
    assert(maxb > 0);
    assert(maxp > 2);
    assert(p_.a > 0);
    assert(p_.x == 3);
    assert(p_.b == 1);
    assert(p_.y == 3);
    p_.y--; // first next() call will be starting point

The following is an iterator over the points. It is specialized for the case of a fixed a value, but can easily be used to generate an entire space, or modified to perform other iteration strategies:

  inline point& next(bool *done) {
    assert(p_.a == a_dim_);
    if (++p_.y > maxp_) {
      p_.y = 3;
      if (++p_.x > maxp_) {
        p_.x = 3;
        for (;;) {
          if (p_.b > p_.a) {

            // this is where b rolls over. when generating the entire space of
            // points this is the point we would increment "a". For example:
            //   p_.b = 1;
            //   if (++p_.a > maxb_) {
            //     *done = true;
            //     return p_;
            //   }
            // Since we only want to iterate over the space for a given "a"
            // value this is the point we return `done = true` to the caller.

            // bump p_.a so we can assert on its uniqueness
            *done = true;
            return p_;

          } else if (gcd(p_.a, p_.b) > 1) {
          } else

    *done = false;
    return p_;

  int maxb_;
  int maxp_;
  struct point p_;
  int a_dim_;

So, that's it. Testing the axby class is fairly basic. For each a value in a range we create a new axby iterator instance:

class TestAxby(unittest.TestCase):
    def test_all_points(self):
        maxb = 200
        maxp = 200
        for a in xrange(1, maxb+1):
            axby = beal.axby(maxb, maxp, a)

Then we iterate over the space in Python and assert that the axby iterator implemented in C++ returns the exact same sequence:

            for b in xrange(1, a+1):
                if fractions.gcd(a, b) > 1:
                for x in xrange(3, maxp+1):
                    for y in xrange(3, maxp+1):
                        done, point = axby.next()
                        assert(not done)
                        aa, xx, bb, yy = point
                        self.assertEqual(a, aa)
                        self.assertEqual(x, xx)
                        self.assertEqual(b, bb)
                        self.assertEqual(y, yy)
            done, point = axby.next()

Putting it all together, we can test an entire state space search.

End-to-end Testing

We'll need something to compare our results to. We've taken the results published at http://www.danvk.org/wp/beals-conjecture/ for a 1000x1000 state space, and also used the published code on that website to generate several other reference data sets. These results are stored in the gold/ directory. First we load all of the gold results into a set:

class TestSearch(unittest.TestCase):
    PRIMES = [4294967291, 4294967279]

    def __get_gold(self, maxb, maxp):
        results = set()
        regex = re.compile(r"(\d+)\^(\d+) \+ (\d+)\^(\d+)")
        filename = "gold/danvk_%dx%d.dat" % (maxb, maxp)
        with open(filename) as f:
            for line in f.readlines():
                m = regex.match(line)
                if not m:
                a = int(m.group(1))
                x = int(m.group(2))
                b = int(m.group(3))
                y = int(m.group(4))
                results.add((a, x, b, y))
        return results

Finally, we use all of our code to perform a state space search one a value at a time, simulating the distribution of the problem. See test.py for checking other problem sizes.

    def __check_gold(self, maxb, maxp, primes):
        results = set()
        search = beal.search(maxb, maxp, primes)
        for a in range(1, maxb+1):
            hits = search.search(a)
        self.assertEqual(results, self.__get_gold(maxb, maxp))

    def test_1000x1000(self):
        self.__check_gold(1000, 1000, self.PRIMES)

Open Questions

  • Are there other ways that the state space can be trimmed?
  • Are there classes of points that are more interesting to test (e.g. large exponents)?
  • Are there methods for making the modulo-based filtering more effective (e.g. larger primes)
  • What improvements can we get by using GPU-based acceleration (e.g. CUDA or OpenCL)?
  • Can we make searching the c^z space even faster by taking advantage of cache locality?

The following discussion about new approaches is quoted from http://norvig.com/beal.html:

Beyond that, we need a new approach. My first idea is to only do a slice of the zr values at a time. This would require >switching from an approach that limits the bases and powers to one that limits the minimum and maximum sum searched for. >That is, we would call something like beal2(10 ** 20, 10 ** 50) to search for all solutions with sum between 1020 and 1050. >The program would build a table of all zr values in that range, and then carefully enumerate x,m,y,n to stay within that >range. One could then take this program and add a protocol like SETI@home where interested people could download the code, >be assigned min and max values to search on, and a large network of people could together search for solutions.

Ultimately, I think that if there are any counterexamples at all, they will probably be found through generating functions >based on elliptical curves, or some other such approach. But that's too much math for me; I'm just trying to apply the >minimal amount of computer programming to look for the "obvious" counterexamples.