Hacker Newsnew | past | comments | ask | show | jobs | submit | amzin's commentslogin

In progress: I’m writing a book about how the brain processes text in general - and news articles in particular.

I also created and maintain a Russian "newspeak" dictionary: https://github.com/alamzin/az/


Professional procrastinator here. All the tips in the post are common knowledge and, sadly, not very good (especially in the long run).

The only thing that works for many people is to skip the motivation part and embrace the rather uncomfortable principle of "action before motivation."

The flow state will come. I believe it arises independently of motivation. Motivation just tricks us into believing that everything we do should bring joy.

It will — but not right now; we need to dive in first.


yeah its sort of the no pain no gain approach but I like it. my best hack is time logging. I feel that if I specifically label breaks or procrastination or production failures, I at least feel like I'm more in control. I like the act of blogging about it though, it never fails to pick me up knowing I'm not alone with suffering from procrastination and I am always happy to read thoughts about it.


Is there a FS that keeps only diffs in clone files? It would be neat


I wondered that too.

If we only have two files, A and its duplicate B with some changes as a diff, this works pretty well. Even if the user deletes A, the OS could just apply the diff to the file on disk, unlink A, and assign B to that file.

But if we have A and two different diffs B1 and B2, then try to delete A, it gets a little murkier. Either you do the above process and recalculate the diff for B2 to make it a diff of B1; or you keep the original A floating around on disk, not linked to any file.

Similarly, if you try to modify A, you'd need to recalculate the diffs for all the duplicates. Alternatively, you could do version tracking and have the duplicate's diffs be on a specific version of A. Then every file would have a chain of diffs stretching back to the original content of the file. Complex but could be useful.

It's certainly an interesting concept but might be more trouble than it's worth.


ZFS does this by de-duplicating at the block level, not the file level. It means you can do what you want without needing to keep track of a chain of differences between files. Note that de-duplication on ZFS has had issues in the past, so there is definitely a trade-off. A newer version of de-duplication sounds interesting, but I don't have any experience with it: https://www.truenas.com/docs/references/zfsdeduplication/


VAST storage does something like this. Unlike how most storage arrays identify the same block by hash and only store it once VAST uses a content aware hash so hashes of similar blocks are also similar. They store a reference block for each unique hash and then when new data comes in and is hashed the most similar block is used to create byte level deltas against. In practice this works extremely well.

https://www.vastdata.com/blog/breaking-data-reduction-trade-...


That’s very interesting. Typically a Rabin fingerprint is used to identify identical chunks of data.

Identifying similar blocks and, maybe sub-rechunking isn’t something I’ve ever considered.


ZFS: "The main benefit of deduplication is that, where appropriate, it can greatly reduce the size of a pool and the disk count and cost. For example, if a server stores files with identical blocks, it could store thousands or even millions of copies for almost no extra disk space." (emphasis added)

https://www.truenas.com/docs/references/zfsdeduplication/


Not an FS but attempting to mimic NTFS, SharePoint does this within it's content database(s).

https://www.microsoft.com/en-us/download/details.aspx?id=397...


That’s how APFS works; it uses delta extents for tracking differences in clones: https://en.wikipedia.org/wiki/Delta_encoding?wprov=sfti1#Var...


APFS shares blocks so only blocks that changed are no longer shared. Since a block is the smallest atomic unit (except maybe an inode) in a FS, that’s the best level of granularity to expect.


With extent-based filesystems you can clone extents and then overwrite one extent and only that becomes unshared.


Sent the link to my son who is an avid gamer.

Got an one-word response:

"Winnienet".

Could not phrase it better.


And to clarify, the beginning is here: http://scp-wiki.wikidot.com/antimemetics-division-hub


> Gitlab has no Chinese or Russian employee so far -> so nobody got pressure or intimidate, ffs.

Actually, they do have employees there. There are at least 5 Russians and 1 Chinese listed here: https://about.gitlab.com/jobs/

There are no Russians at the role that had been debated though.


Came to mention Notion.so too.

My email is a mess for I should switch between several work/home mailboxes — and from time to time I just have to wait for an email too.

This is one of the reasons I avoid launching Notion.so.


Yandex.Factory is not a fund, actually. It is investment program.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: