Ancestris, Version 11 Notes

Ancestris 11, an update from the previously reviewed version 10, is much more robust now. As a major improvement, the development team put up comprehensive documentation on their site.

After having gotten used to the interface, which is customizable, things have improved. Navigating the three editor panes is easy, just pick the one you like and go with it.

The output options are still bare-bones web page output. I’d like to see something like a register-style report than have whatever I’ve chosen to “print out” be put up on the internet. It’s still easier to have Lifelines do the report for me. I wrote a modified version of one that came with Lifelines to mimic the New England Historical Genealogical Society’s format.

Ancestris is a free genealogy program written in Java. It runs on Windows, Macintosh, and Linux systems.

NPM

An Update on GenealogicNG

I’ve just completed a very basic, rudimentary implementation of the database in Rust. Rust is an open-source programming language and will be used for future development of GenealogicNG. Right now, it’s only the CRUD (create, read, update, and delete) functions that are usable. There is no interface, fancy, or otherwise. 🙂

The code is available at GitHub (https://github.com/npmaling/genealogicng-code).

The background materials are also at GitHub (https://github.com/npmaling/genealogicng)

Also, here are a few posts from the past on this site:

https://seagenesfhgr.com/2013/03/07/physical-version-of-the-genealogical-data-model/

https://seagenesfhgr.com/2013/02/28/an-introduction-to-genealogicng/

.

NPM

What could be better? Books? Genealogy? Both!

I’ve been watching Just Genealogy on YouTube and it is fascinating. Craig Scott, a prominent genealogist hosts short, highly informative clips on genealogy and books.

Mr. Scott is proprietor of Heritage Books, a long-time fixture in the history and genealogy business. He has branched out to YouTube to present new books in his stock and discuss them with viewers. He also accepts viewer questions to address such issues as “I’ve looked at all the records, now what?”

His regular clips on the genealogy standards put out by the Board for Certification of Genealogists are also interesting. He analyzes and comments on the basis for each standard in the revised second edition of the guide, here.

This is an engaging series that I’ll keep an eye on.

Thanks, Mr. Scott!

NPM

Genealogy as a Form of Data Analysis

This paper is loosely based on the Wikipedia article entitled “Data Analysis” and the book Mastering Genealogical Proof.

Genealogists use raw data to accumulate and analyze patterns and trends toward establishing a Genealogical Proof. Evidence in the genealogical community is generally understood as pieces of data that are arranged through collection, sifting, and arranging. Evidence, positive or negative, is acquired through examining and modeling data using generally accepted processes. One such process is to use the computer application Evidentia. Other processes enabling the development of evidence are those used in the legal and forensics professions (e.g., DNA analysis).

Each point of data genealogists use is inspected, cleansed, transformed, and modelled. Most serious genealogists use the Genealogical Proof Standard. While this standard is more qualitative than quantitative, the results are the same, actionable information used to formulate decisions.

The Genealogical Proof Standard follows, simply: Formulating a research question, gathering data sources, considering the information in those sources, formulating evidence from that information, and finally constructing a proof statement. The process is generally iterative since there is no such thing as a final statement of proof in genealogy.

While traditional data analysis is generally thought to be quantitative, there is much similarity to the genealogical research process. The steps in data analysis are analogous to the process used by genealogy professionals. Data analysis begins with a research question, followed by compiling source information, and finally, generating actionable conclusions.

Research Question

Sometimes thought of as a hypothesis, the research question is the beginning of both genealogical research and data analysis. Genealogists formulate a question by asking something such as “Who was Joan Jones’ mother?” Data analysts ask, “How is product A better than product B?” The answers come in basically the same way for both.

Data Collection, Processing, and Cleaning

To answer the research question, both genealogists and data analysts collect, process and classify data relevant to the issue. Almost all data is seen as relevant to analysts, but genealogists often go further, collecting source material relevant not only to the issue, but also surrounding the issue. Data analysts, on the other hand, are more focused on the question itself, locating only data relevant to products A and B.

The difference between traditional and genealogical data analysis is that genealogists have much more fuzzy information to deal with. Items like local and regional history books may include data about their question. Such items are generally not relevant to a data analyst focused on a product research project, unless it involves cultural appropriation, i.e., the Korean car makers’ KIA Tucson vehicle. 😊

Exploratory Analysis

Genealogists often explore different sets of data to glean information and evidence relevant to their questions. Similarly, a traditional data analyst will do the same, focusing more on specific items than general items.

Modelling and Algorithms

There are no “real” algorithms for genealogists to apply to their data findings. There is, however, a Genealogical Data Model, which was constructed to help genealogists apply their data to real-world projects. The Genealogical Data Model was originally constructed to be a basis for software, but since it was completed, no software has used the GDM (except for The Master Genealogist, which used large parts of it).

Data Products and Communications

Genealogists use a proof model to present data and their formulation of the evidence they’ve compiled. A traditional data analyst uses a tool such as business intelligence software to present their findings. The only real difference between the two is that they present findings in a different way.

NPM

Fascinating New Find About Richard Mellen

Just found a newish website whose author is apparently unwilling to commit to scholarly diligence and credibility. Doug Sinclair wrote a paper about Richard Mellen of Massachusetts, alleged father of Simon Mellen of Sherborne and Framingham.

Mr. Sinclair states that his website doesn’t claim to the standard of scholarly journals, and it certainly doesn’t.

Further, he ignores the clear, cogent, and concise statements at the beginning of my paper on Richard Mellen, and the genealogy of Simon Mellen, that they are each “extended literature review[s].” Further, the next paragraph in those publications cogently states that the focus of each was on “published record sets.” What is not credible or clear about that?

Mr. Sinclair fails to cite his sources for the statements concerning my work, yet appears to cite everything else he writes about. Nothing new was presented that hadn’t been already written about by myself or others, just pictures. One picture in particular clearly shows he read my blog article and paper about Richard “Maling” aka “Waling.” The picture he posted clearly shows a “W”.

Any diligent genealogical researcher will use both of our works as clues only. Any diligent genealogical researcher will also follow the BCG Code of Ethics in using our works.

NPM