/r/CouchDB

Photograph via snooOG

/r/CouchDB

1,076 Subscribers

1

Need Help Please Deploying CouchDB

Hi guys! I need some help for deploying my code to railway, especially with my database usong CouchDB.

TDLR: for CouchDB, what should I set my variable name as well as for my reference as in my empty project on railway?

My project backend is using springboot, and for databases, I have SQL, MongoDB, and CouchDB as my databases.

So I created an empty project to deploy my backend project. Inside, I have to create 3 environmental variables. When deploying in railway, under my project variables, there will be 3 variables: one for SQL, MongdoDB and also CouchDB. For railway, after adding SQL and MongoDB, environmental variables for URL is automatically added, but this is not the case for CouchDB!

So for my empty project under its environmental variables, I put this: SPRING_DATASOURCE_URL under my variable name and for reference I put this: jdbc:${{MySQL.MYSQL_URL}}. This is for SQL.

I put this: SPRING_DATA_MONGODB_URI under my variable name and for reference I put this: ${{MongoDB.MONGO_URL}}/products?authSource=admin. This is for MongoDB.

Then for CouchDB, what should I set my variable name as well as for my reference as?

Thank you very much!

0 Comments
2024/03/17
16:00 UTC

2

Migrating data from Firestore to PouchDB

Hey all,

I am adding offline capabilities to my app and decided to move away from Firebase even though Firestore offers some kind of persistence. I found PouchDB which seems perfect for what I need, I just needed some advice for the current database structure I have in Firestore.

Basically there are 2 main collections: "users" & "projects". A user can have multiple projects. So far so good. My issue is that each project document has 4 more sub-collections, which as I understand PouchDB doesn't support? I don't expect these sub-collections to have more than a few dozen documents each, so perhaps I could just add them as keys to each project but on the other hand I don't always need this data when I'm fetching a project.

I'm not a database expert so I'm wondering if there's a better approach? Any help is appreciated.

0 Comments
2024/03/14
03:49 UTC

1

Expose couchdb using cloudflared tunnel

I am trying to configure the couchdb to be accessible from internet using the cloudflared dns.

I started the cloudflared tunnel using following config

  - hostname: example.org
 path: /couchdb
 service: http://192.168.0.103:5984

so that it is accessible via https://example.org/couchdb

but upon visiting the url I am faced with the following error on the webpage, and 404 error in the couchdb docker logs.

{"error":"not_found","reason":"Database does not exist."}

1 Comment
2024/03/11
05:50 UTC

2

Recliner-JS: A local REST CouchDB in your browser

Recliner-JS

A CouchDB like DB, which runs in your browser. Access your saved attachments locally with REST API:

<img src=`/recliner/dbname/doc_id/attachment_name`>

GITHUB_LINK

Features

  1. Uses same REST API like Couch DB.
  2. Implements CouchDB replicator protocol for seem less replication with backend.
  3. Can Lazy load Blobs(video media) from cloud.
  4. Partial content and stream supported.
  5. Uses IndexedDB as its backend DB and hence no storage restrictions.
  6. Mango like queries for searching and replication.
  7. Typescript based client to access REST API.

Usage

In your service worker add this:

import {Recliner} from 'recliner-js';
const recliner = new Recliner();//create instance
self.addEventListener("fetch",(e)=>{
    const url_path = Recliner.getPathFromUrl(e.request);
    //mounts recliner
    if(url_path.startsWith("/recliner")){
        e.respondWith(recliner.process(e.request));
    }else{
        // do whatever else
    }
});

Now you can access docs and attachments saved in your recliner DB via URL

<img src=`/recliner/dbname/doc_id/attachment_name`>

CRUD with DB

There are two ways to interact with DB:

  1. Using regular fetch in your JS code using REST API similar to CouchDB*.

const getADoc = await fetch(`/recliner/dbname/docid`);
if(getADoc.status === 200){
    return await getADoc.json();//{_id,_rev,ok: true}
}

See complete list of Rest API

*Though many , but not all Couch REST API is supported. See Difference from CouchDB section.

  1. Use the a client instead: UsageDAO

    import {UsageDAO} from 'recliner-js'; //Create await UsageDAO.postADoc(dbname,{name:"person1",age:30}); //Retrieve const doc = await UsageDAO.readADoc(dbname,docid); //Update await UsageDAO.updateADoc(dbname,doc._id,{name:"person1",age:34}); //Delete await UsageDAO.deleteADoc(dbname,doc._id);

    //query const findResult = await UsageDOA.findByPagination({ dbanme, selector:{ age:{$lt: 40}, income: {$within:[10000,20000]}, loc: {$isinpolygon:[]}//has some GIS capability } });

    //Save Attachments await UsageDAO.addAttachmentsToDocID(dbname,doc._id,{ "my_doc.pdf":DOC_BLOB, "my_video.webm":VIDEO_BLOB
    });

    //Save attach wth a cloud URl. //this way when such docs get replicated the Attachments are not sent. As they can be downloaded at end system using the cloud URL await UsageDAO.addAnAttachmentToExistingDoc(dbname,doc,attachment_name,blob,new_edits,cloud_url,content_type); //CRUD with attachments on DOC is available

    //Replication: say fetch last 10 post await UsageDAO.replicate({ selector:{ post_id type:"post", time:{$lt: now} }, limit: 10, target:{ url:"/recliner/my_db" }, source:{ url: "/proxy/couchdb/dbname", headers:{ token:"some_token" } } });

Partial content and Media streaming

Save the document with a cloud_url in a _attachments property.

await fetch("/recliner/dbname/docid",{
    method:"PUT",
    body:{
        _id:docid
        name:"person1",
        _attachments:{
            "my_video.webm":{
                cloud_url:"some_valid_cloud_url_which_supports_partial_content",
            }
        }
    }
});

Now this is can be streamed using:

<video src="/recliner/dbname/docid/my_video.webm">

The video player will automatically stream the video via recliner. Using the cloud_url, the docs will be partially downloaded and saved for offline use, and then streamed to video element. So next time when user stream the same video, its pulled out from the local cache.

However for all this to work, you need to configure recliner for what all mime type you want to support for streaming.

import {Recliner} from 'recliner-js';

const recliner = new Recliner(24,{
    "video/webm":1000_000,//1MB of partial content size for streaming
    "audio/mp3":1000_00//0.1MB of partial content size for streaming
});

When configured this way, then whenever an attachments of type webm and mp3 are requested, they are automatically streamed. If partial content of a doc is not present locally, than using the cloud_url its partial content is first pulled from cloud, saved in indexedDB and then streamed to the corresponding requesting GUI components like : video and audio tags. Next time same partial content will be streamed from local DB, instead of fetching it from cloud_url.

REST API supported

"/recliner/:db";
"/recliner/:db/_design/:ddoc";
"/recliner/:db/_find";
"/recliner/:db/_index";
"/recliner/_replicate";
"/recliner/:db/_changes";
"/recliner/:db/_bulk_get";
"/recliner/:db/_revs_diff";
"/recliner/:db/_local/:docid";
"/recliner/:db/:docid";
"/recliner/:db/:docid/:attachment";
"/recliner/:db/_db_design";
"/recliner/:db/_run_update_function";
"/recliner/_delete_recliner";

Gotchas

  1. For a multientry search the field name muts end with _m
  2. Supported query operators: $lt,$lte,$eq,$ne,$gte,$gt,$exists,$within,$nin,$regex,$in,$isinpolygon,$isnotinpolygon,$nwithin

Difference from CouchDB

  1. DB level design docs, saved via UsageDAO.putADBDesign, can be used to configure various function at DB level :

export interface DBDesignDoc{
    //name of the DB
    for_db:string;
    //before insertion docs are validated using this 
    doc_validation_function?:string;//this is stringified JS function
    
    //used for map reduce
    map_functions?: Record<string,string>;
    reduce_functions?:Record<string,string>;

    //can be used to mass modify docs using a selector
    update_functions?:Record<string,string>;

    /**
     * Used during remote to local replication using view query as source. [one can pass viewQueryUrl to replication info, to start view based replication]
     * This functions are used to filter view result before replicating them to local DB.
     */
    view_result_filter_functions?:Record<string,string>;
}
  1. Views are not supported, however Indexes, Map and Reduce is still supported using UsageDAO.postQueryToDBDesign<D>(dbname:string, query:MapReduceQuery). 3. Design docs are not replicated by default when doing remote to local or local to remote replication. However for local to local replication design docs are copied. By local means database present in the browser. By remote means the one located and accessed via HTTPS on DB server(or via a proxy). 4. _local attribute can be added on doc. They remain solely on local machine, and is removed when are replicated. So some values you wanted to keep in doc, but don;t want to send to server can be saved here. CouchDB Local Docs is supported, which are never replicated.

Though the most important API, to deal with docs and attachments and Database is implemented. And many which deemed insignificant where dropped.

Running the Demo

Type this commands in order

  1. npm run i install dev dependencies
  2. npm run build builds for typescript system
  3. npm run predemo adds ".js" extension to build for demo purpose.
  4. npm run demo : open http://localhost:8000/demo/index.html to view the demo
1 Comment
2024/03/09
14:12 UTC

2

Illegal hostnames and k8s statefulsets

I’m not finding documentation that coherently explains a lot of stuff in couchdb. Seems to be an ongoing problem throughout its 10+ year existence according to what shows up in search engines, but my question today is: what is a legal hostname in couchdb? Is it just letters, numbers, and period - or are “hyphens” considered legal?

K8s has something called a “statefulset” that (long story short) auto generates hostnames with a hyphen followed by an ordinal. The FQDN would look something like “couchdb-0.couchdb.apache-couchdb.svc.cluster.local”, and scaling would create nodes couchdb-1, couchdb-2, etc… same FQDN pathing. Unfortunately, couchdb@couchdb-0.couchdb.apache-couchdb.svc.cluster.local throws the illegal nodename error.

Other that creating separate deployments with separate non-hyphenated service names (seems kind of silly), there’s no way to set this up without a whole lot of kludgy workarounds - unless you know how to bypass this hyphen issue.

0 Comments
2024/01/09
21:02 UTC

2

Comparing two field in a document in CouchDB

I am currently working on a project that uses CouchDB, and I need to write a Mango Query, in which two fields needs to be compared, and I was not able to find too much about this online, and didn't have luck figuring this myself.

Long story, short, I have documents which contain the following fields, among many others: "processed" and "modifiedOn". These fields are of type string and their content is date formatted with ISO-8601 date format (yyyy-MM-ddThh:mm:ss).

ex:

{
  ...,
  "processed": "2023-12-12T12:12:12Z",
  "modifiedOn": "2023-12-12T12:13:00Z"
}

So, my question is whether a query can be written, so that it will return all of the documents whose "processed" field's value is less that the "modifiedOn" field's value.

1 Comment
2023/12/12
14:40 UTC

3

CouchdDb cluster, replication and load balance

Hello, I have been using couchdb in standalone mode for some time and I am considering setting up a cluster for reasons of redundancy and also load balancing. As I have read, couchdb has its own load balancing system that, although it is quite simple, is sufficient for the test I want to do.

I have created a cluster of 3 nodes in Docker and the 3 communicate and replicate well, however, no matter how many queries I launch to node 0, (millions in 1 minute), I do not see that it delegates any to the rest of the nodes.. Should I configure something else or I have not understood couchdb balancing?

Thanks people.

2 Comments
2023/10/14
16:24 UTC

2

Constant CPU usage in Docker

Hi, is it normal for the Docker container's CPU usage to consistently hover around 5%? I've noticed that beam.smp seems to be responsible for this, at least according to what I see in htop. Typically, this wouldn't be a concern, but I'm worried about unnecessary power consumption. I have several other services running, and this behavior is only occurring with CouchDB.

1 Comment
2023/09/27
07:06 UTC

1

newby question - 409 error

I am new to couchDB. I inheritted a couchDB server at work. I am looking through it and at times we are getting 409 errors. I know that is a document conflict. My question is is there a way to know why that is happening? Is it because I am attempting to update an old revision? Is that the only way that a 409 error is generated?

2 Comments
2023/09/02
00:30 UTC

3

CouchDB repeatable performance testing

How well does CouchDB perform with 100,000 per-user databases supporting "live" users? Are there any readily available and repeatable online tests to gauge its performance?

I find the implementation of CouchDB to be laborious, time-consuming, and not enjoyable. I'm concerned that the advantages of replication might not justify the effort and inconvenience involved in using CouchDB.

6 Comments
2023/07/28
22:13 UTC

3

Securing expose couchdb (on www)

Hello,

I'm pretty new in couchdb world, I just use it to synchronize obsidian (with livesync plugin) but I wonder what is best practices to securing a couchdb exposé on web. Is use this configuration in docker :

[couchdb]
single_node=true
max_document_size = 50000000

[chttpd]
require_valid_user = true
max_http_request_size = 4294967296

[chttpd_auth]
require_valid_user = true
authentication_redirect = /_utils/session.html

[httpd]
WWW-Authenticate = Basic realm="couchdb"
enable_cors = true

[cors]
origins = app://obsidian.md,capacitor://localhost,http://localhost
credentials = true
headers = accept, authorization, content-type, origin, referer
methods = GET, PUT, POST, HEAD, DELETE
max_age = 3600

It's behind a reverse proxy in https (manage by cloudflare), password it's secure (32 chars with upper, lower and number).

But I wonder if it's enough? I read official documentation but I found nothing else than require_valid_user and use strong password.

Do you have recommandation ?

Thank on advance

3 Comments
2023/06/17
14:36 UTC

0

Bye Reddit! Remember to delete your posts before deleting your account!

Your value to Reddit is your free posts and comments so remember to delete your posts before deleting your account!

I used this Reddit API script to delete 10 years of my comments and posts: https://codepen.io/j0be/full/WMBWOW/

Bye Reddit! It has been fun!

1 Comment
2023/06/12
00:32 UTC

2

Should I store all object attributes of type object in one document?

I am just starting up with CouchDB. Say I have an object joeDoe of class Person that has an ArrayList attribute of Car objects. My question is this: Should I store my joeDoe and all his cars in one document, or should I store each object in separate documents and simply provide the ids of the objects wherever they are needed? Which is considered best practice? Is there any benefit or drawback in any of these approaches?

Thank you.

2 Comments
2023/05/14
16:40 UTC

0

Problem at creating view document

Dear CouchDB users,

I downloaded CouchDB to a MacBook Pro from this address "https://neighbourhood.ie/download-apache-couchdb-mac/" and installed it. I was trying to create a view document but I was taking "Error: internal_server_error message". So I just tried to verify the installation and took this screenshot. What seems to be the problem?

https://preview.redd.it/lqpl77r4kuqa1.png?width=2540&format=png&auto=webp&s=c810a277460fd9f0bdf40a75418963d21f304647

0 Comments
2023/03/30
09:41 UTC

7

Announcing my new realtime database, Fireproof

Hello, I'm one of the creators of CouchDB and PouchDB, and I have a new database people in this forum might be excited about. It's free and open source, and uses the IPFS protocol. The API feels a lot like CouchDB, except this is writtin in (not very much) JavaScript and designed to run in pages and app. I'd love feedback on ease of use and API from folks in this community -- I bet y'all are better qualified than most to understand how to use something like this. There are a bunch of opportunities to contribute listed in the README, wow that's be exciting if I saw some PRs.

Here's the Fireproof GitHub repo, and here's the website. Thanks! Chris

0 Comments
2023/03/08
04:49 UTC

2

Yet Another Database Design question (pouchdb and couchdb)

Hi

I know this has been asked and answered a few times, but I'm going to ask again, because I'm still unsure.

(Sorry - this has ended up being quite long - TLDR: one database per user sounds great for offline PouchDB stuff - but how do you make it work when multi-user access to shared documents with fine-grained permissions is needed?)

I've got a V1 app, written in Rails using a relational back-end. I'm now approaching the time to design the V2 version and the client wants it to be able to work offline (which immediately puts Rails out of the question - at least for the client, if not the admin interface). PouchDB and CouchDB seem like the perfect way to do this - but my relational mind is still struggling to figure out how to organise things. Documents and Views I get - but fine-grained security and authorisation less so.

In Rails all client access to the data is through the app-server, so I control who sees, edits and deletes which document. But if the system is to work offline, my PouchDB database needs to sync to the server-side CouchDB database, bypassing any app-server level controls.

Each user only has access to a subset of the data - so I don't want to sync the entire database across. Firstly, it's costly (Gbs to move) and secondly, I don't want people poking around on their client device and seeing other people's stuff inside the database (even if they can't access it in the app - the client has some security-conscious customers).

"One database per user" seems to be the solution - but a lot of this data is shared. For example (and this is just a small subset) - a supervisor creates a work-schedule, it gets approved by a manager, and then the employee views it. When it's time to start working, the employee updates their timesheet. The timesheet gets submitted back to the supervisor and eventually processed by the manager.

The account owner sees/updates everything across all departments. The manager sees/updates everything within their own department. The supervisor only sees/updates the schedules and timesheets for their own team. The employee only sees/updates their own stuff.

My initial thought, then, is to have a primary database, then a database per user. Then, I set up replication filters between all these databases so the correct information goes to the correct place - in both directions. Does that sound like a good idea?

(Even more complex - when not just dealing with timesheets, certain types of document might need to be available to be visible to and edited by employee-1, then visible to and edited by employee-2 - so the filter rules would have to allow updates from employee-1-database to primary to employee-2-database and back again)

Then within each document (schedule, timesheet etc), on the primary I have a list of users who have access to it, so the filter rules can easily figure out who can see it? Although that then potentially publishes a list of all users to the user-databases. So can the filter rule transform the document in some way? Or can the filter rule reference a separate document which describes the authorisation rules for this document?

Finally when they sign up a new employee I have to create a new database (which will be a standard template, with filter rules predefined, so should be pretty simple) and then possibly add in extra filter rules to the replication design document on the primary database (depending on how the permissions are stored)? Likewise, if someone gets promoted, from supervisor to manager, I then need to rewrite the filter rules relating to them, both on their user-database and on the primary?

Or is there another simpler method that I'm missing?

5 Comments
2023/02/23
07:52 UTC

2

PouchDB and creating the database with default docs

Hi all, I'm playing with PouchDB for an offline-first web app and wondering how best to solve a simple thing.

What I want is to be able to create the database with a couple of default example docs. There doesn't seem to be any obvious response from `new PouchDB('example')` that would tell me 'this is newly created' that I can use to trigger the default document creation.

I could put a flag somewhere, maybe a 'config' document with 'hasInitialised' in it and use that. But it seems a bit of a faff to create a whole additional database with a single document in it to store my config for a single flag. Is there something obvious I'm not seeing?

Thanks!

2 Comments
2023/01/16
15:34 UTC

4

How to create a running total in CouchDB?

Assume db holds bank account transactions, each transaction have a date and amount. How to create a view that will give transactions sorted by date, amount and running total (sum with previous amount incrementally)?

map only works with a single document, no access to the previous document. reduce reduces the grouped items by some aggregation. No idea how to access the previous row.

This is called a Window Function in SQL databases. Any help is appreciated.

1 Comment
2022/12/27
06:47 UTC

2

is CouchDB only suitable for very tiny databases?

I installed CouchDB on a cloud server instance w/ 512MB RAM, 20GB disk, and uploaded 200,000 json documents, totaling just under 1GB of documents.

Then I tried to create a simple view (conditional `emit` of 2 fields).

During the view creation I got "OS timeout".

Then trying to use the view I get "OS error 137".

(these are from memory as the error pop-up in Fauxton goes away before I could copy/paste)

Is this normal?

6 Comments
2022/12/01
22:04 UTC

9

I tried CouchDB and I really like it

My latest blog entry on this here. I am mainly comparing it with MongoDB as far as NOSQL systems is concerned.

0 Comments
2022/11/18
16:21 UTC

3

how do you develop complex views? writing javascript inside an escaped double quoted string doesn't seem sustainable beyond simple Hello World examples

Those of you who are experienced CouchDB devs writing some very complex views, what does your IDE and test/deploy pipeline look like?

1 Comment
2022/11/14
01:56 UTC

1

Does CouchDB really keep a whole replication of the database on the client-side too?

I'm new to CouchDB, and more generally I'm new to "offline-first".

I read somewhere that databases such as CouchDB that make offline-first possible, store a replication of the whole user's dataset on their own side (client-side) as well, is it true?

If that's the case, it doesn't make sense to me... Please tell me what I'm getting wrong here? It means every time the user logs in to their account on a new device, or even on a new browser, (or worse, when Incognito browsing is used), their data must be downloaded fully? Couldn't it be chopped into parts and then -like pagination-, keep only a more recent part on the client-side, then load more on need?

Of course, the new device/browser thing I said doesn't happen frequently (the Incognito thing could happen more frequently though), but even if these happen every few times in a while, can be UX killer... Let's say it's a simple notes app with only 5000 notes, it will be almost 50MBs in size to download for the first time, that means a noticeable delay even on a good Internet connections before the initial UI load... Isn't there a way to make this experience more smooth?

6 Comments
2022/11/10
07:00 UTC

2

Looking for Flask example

Can you recommend a simple flask example to use CouchDB from Flask ?

I am newbie for both of them.

0 Comments
2022/07/19
08:29 UTC

2

Which free and opensource NoSQL database provides feature for creating group/bucket of documents?

I am learning CouchDB. As I understand it, documents in the database cannot be grouped into categories, such as, for example, all receipt documents can be put into a receipt bucket, invoices can be put into invoice bucket etc.

Are there any free and opensource NoSQL databases that provide this feature of grouping documents according to category?

1 Comment
2022/05/27
15:03 UTC

2

Settings issue

Anyone know what causes this error when trying to view the settings in Fauxton?

Failed to load the configuration. Unexpected token < in JSON at position 0

2 Comments
2022/05/12
20:30 UTC

2

Post/Comment DB design: Postgresql v/s CouchDB

I am comparing DB design for a simple "Post and Comment" system using Postgres and CouchDB. With Postgres I can design the following tables:

user_info {email, pass_hash, pass_salt, ...}

post_info {post_id, creator_email, title, text, ...}

comment_info {comment_id, creator_email, post_id, parent_comment_id, text, ...}

But if I use CouchDB, there is a concept of creating per-user tables. So I was thinking of the following design:

user_table {email, table_id}

user_<table_id> {email, pass_hash, pass_salt, ...}

post_<table_id> {post_id, <table_id>_creator_email, title, text, ...}

comment_<table_id> {comment_id, <table_id>_creator_email, <table_id>_post_id, <table_id>_parent_comment_id, text, ...}

I am in no way expert in Postgres and CouchDB, so my question is, is this the correct way to design per-user CouchDB tables? What is the better way? And what is the efficient way to create/use CRUD queries?

2 Comments
2022/03/28
14:09 UTC

2

It's it possible to connect CouchDB to ldap for auth?

I've seen one plugin that hasn't been updated in 6 years, and Google thinks I want to use CouchDB as a DB for ldap.

All I want is to have CouchDB authenticate against an ldap service like IPA.

Can someone point me in the right direction? Many thanks

1 Comment
2021/12/16
23:44 UTC

4

CouchDB is terrible if not dysfunctional for large documents

I've been forced to come to this conclusion after getting a ton of timeouts and 500 errors when trying to simply replicate a database that contains a few 300MB JSON documents.

Querying Mango is also a futile exercise, it just times out.

I managed to resolve one issue, which was the system pulling the replication outputting the following in the log:

[error] 2021-12-08T22:19:49.862114Z couchdb@127.0.0.1 <0.22598.2> -------- Replicator, request GET to "http://localhost:5999/invoices/_changes?filter=filters%2Fdeletedfilter&feed=normal&style=all_docs&since=%222795484-g1ABjnwushUHF4iUF87asdf72hj3lkj4lkj28sdfd8&&Fikjsdlkjjr___-IJ2349sjdfglkjOLIJlk34l2kj3ijlIJFasdf_zjaihuHYUFhw;kljsdj442kjla9s8fqkjf%22&timeout=10000" failed due to error {error,req_timedout}

[error] 2021-12-08T22:17:42.538990Z couchdb@127.0.0.1 <0.2043.0> -------- Replicator, request GET to "http://localhost:5999/invoices/_changes?filter=filters%2Fdeletedfilter&feed=normal&style=all_docs&since=%222795484-g1ABjnwushUHF4iUF87asdf72hj3lkj4lkj28sdfd8&&Fikjsdlkjjr___-IJ2349sjdfglkjOLIJlk34l2kj3ijlIJFasdf_zjaihuHYUFhw;kljsdj442kjla9s8fqkjf%22&timeout=10000" failed due to error {connection_closed,mid_stream}

That &timeout=10000 is 1/3rd the value of the following parameter in /opt/couchdb/etc/local.ini:

[replicator]
connection_timeout = 30000

So I simply added another zero to make the timeout 100 seconds instead of 10.

But now I was getting 500 errors:

[error] 2021-12-08T23:03:42.464170Z couchdb@127.0.0.1 <0.626.0> -------- Replicator, request GET to "http://localhost:5999/invoices/_changes? filter=filters%2Fdeletedfilter&feed=normal&style=all_docs&since=%222795484-g1ABjnwushUHF4iUF87asdf72hj3lkj4lkj28sdfd8&&Fikjsdlkjjr___-IJ2349sjdfglkjOLIJlk34l2kj3ijlIJFasdf_zjaihuHYUFhw;kljsdj442kjla9s8fqkjf%22&timeout=100000" failed. The received HTTP error code is 500

It's now the server holding the original database I'm replicating off that's throwing errors.

[info] 2021-12-08T23:03:42.451681Z couchdb@127.0.0.1 <0.255.0> -------- couch_proc_manager <0.15833.2> died normal
[error] 2021-12-08T23:03:42.451742Z couchdb@127.0.0.1 <0.21493.1> 455997af04 OS Process Error <0.15833.2> :: {os_process_error,{exit_status,1}}
[error] 2021-12-08T23:03:42.451923Z couchdb@127.0.0.1 <0.21493.1> 455997af04 rexi_server: from: couchdb@127.0.0.1(<0.15895.1>) mfa: fabric_rpc:changes/3 throw:{os_process_error,{exit_status,1}} [{couch_os_process,prompt,2,[{file,"src/couch_os_process.erl"},{line,59}]},{couch_query_servers,proc_prompt,2,[{file,"src/couch_query_servers.erl"},{line,536}]},{couch_query_servers,with_ddoc_proc,2,[{file,"src/couch_query_servers.erl"},{line,526}]},{couch_query_servers,filter_docs_int,4,[{file,"src/couch_query_servers.erl"},{line,510}]},{lists,flatmap,2,[{file,"lists.erl"},{line,1250}]},{couch_query_servers,filter_docs,5,[{file,"src/couch_query_servers.erl"},{line,506}]},{couch_changes,filter,3,[{file,"src/couch_changes.erl"},{line,244}]},{fabric_rpc,changes_enumerator,2,[{file,"src/fabric_rpc.erl"},{line,517}]}]
[notice] 2021-12-08T23:03:42.453155Z couchdb@127.0.0.1 <0.15304.1> 455997af04 localhost:5999 127.0.0.1 admin GET /invoices/_changes?filter=filters%2Fdeletedfilter&feed=normal&style=all_docs&since=%222796340-g1AAAACheJzLYWBgYMpgTmEQTM4vTc5ISXIwNDLXMwBCwxyQVCJDUv3___-zMpiTGEQj5ucCxdiNzcyTUgwMsenBY1IeC5BkaABS_-EGBk0FG2iSam5pkpSMTWsWADLTKlk%22&timeout=100000 500 ok 21392

So at this point I give up. I've tried increasing OS process timeouts, fabric timeouts, but... it's so very unfortunate.

CouchDB is supposed to be able to handle 4GB JSON documents. It simply can't. It can't even handle a 200MB JSON document. Even if it could there's zero documentation about how to give CouchDB whatever resources or time it needs to handle such a large document.

4 Comments
2021/12/09
01:24 UTC

Back To Top