Found it. Hyrum said 207 tests were failing a few days ago. Ah well...
Time to fix things.
On Sun, Mar 1, 2009 at 17:14, Greg Stein <gstein_at_gmail.com> wrote:
> After this change:
>
> Summary of test results:
> 819 tests PASSED
> 24 tests SKIPPED
> 24 tests XFAILED
> 279 tests FAILED
> 2 tests XPASSED
>
> I believe that is a slight regression from before my change, but I
> can't find the prior numbers. The change moves the code in the right
> trajectory, but may require some further tweaking to bring the tests
> back in line.
>
> Cheers,
> -g
>
> On Sun, Mar 1, 2009 at 16:05, Greg Stein <gstein_at_gmail.com> wrote:
>> Author: gstein
>> Date: Sun Mar 1 07:05:28 2009
>> New Revision: 36217
>>
>> Log:
>> Rebuild read_entries() around the db_read_children() and db_read_info()
>> interfaces. It still peeks into the database on its own, but only for some
>> very limited information that isn't exposed by wc_db (yet?).
>>
>> * subversion/libsvn_wc/entries.c:
>> (find_working_add_entry_url_stuffs): removed. wc_db does this now.
>> (read_entries): rebuilt. primarily, the second loop over the WORKING
>> nodes is no longer performed since we now have those children in our
>> list (from read_children). there are still quite a few hacks and other
>> oddities to overcome expectations/impedance between the entries and
>> wc_db interfaces, but those will (hopefully) disappear over time.
>>
>> Modified:
>> trunk/subversion/libsvn_wc/entries.c
>>
>> Modified: trunk/subversion/libsvn_wc/entries.c
>> URL: http://svn.collab.net/viewvc/svn/trunk/subversion/libsvn_wc/entries.c?pathrev=36217&r1=36216&r2=36217
>> ==============================================================================
>> --- trunk/subversion/libsvn_wc/entries.c Sun Mar 1 06:48:44 2009 (r36216)
>> +++ trunk/subversion/libsvn_wc/entries.c Sun Mar 1 07:05:28 2009 (r36217)
>> @@ -856,60 +856,6 @@ fetch_wc_id(apr_int64_t *wc_id, svn_sqli
>> }
>>
>>
>> -/* This function exists for one purpose: to find the expected future url of
>> - an entry which is schedule-add. In a centralized metadata storage
>> - situation, this is pretty easy, but in the current one-db-per-.svn scenario,
>> - we need to jump through some hoops, so here it is. */
>> -static svn_error_t *
>> -find_working_add_entry_url_stuffs(const char *adm_access_path,
>> - svn_wc_entry_t *entry,
>> - const char *relative_path,
>> - apr_pool_t *result_pool,
>> - apr_pool_t *scratch_pool)
>> -{
>> - const char *wc_db_path = db_path(adm_access_path, scratch_pool);
>> - svn_sqlite__stmt_t *stmt;
>> - svn_boolean_t have_row;
>> - svn_sqlite__db_t *wc_db;
>> -
>> - /* Open parent database. */
>> - SVN_ERR(svn_sqlite__open(&wc_db, wc_db_path, svn_sqlite__mode_readonly,
>> - statements,
>> - SVN_WC__VERSION_EXPERIMENTAL, upgrade_sql,
>> - scratch_pool, scratch_pool));
>> -
>> - /* Check to see if a base_node exists for the directory. */
>> - SVN_ERR(svn_sqlite__get_statement(&stmt, wc_db,
>> - STMT_SELECT_BASE_NODE_BY_RELPATH));
>> - SVN_ERR(svn_sqlite__bindf(stmt, "s", SVN_WC_ENTRY_THIS_DIR));
>> - SVN_ERR(svn_sqlite__step(&have_row, stmt));
>> -
>> - /* If so, cat the url with the existing relative path, put that in
>> - entry->url and return. */
>> - if (have_row)
>> - {
>> - const char *base = svn_sqlite__column_text(stmt, 0, NULL);
>> -
>> - entry->repos = svn_sqlite__column_text(stmt, 1, result_pool);
>> - entry->uuid = svn_sqlite__column_text(stmt, 2, result_pool);
>> - entry->url = svn_path_join_many(result_pool, entry->repos, base,
>> - relative_path, NULL);
>> - return svn_sqlite__reset(stmt);
>> - }
>> - SVN_ERR(svn_sqlite__reset(stmt));
>> -
>> - /* If not, move a path segement from adm_access_path to relative_path and
>> - recurse. */
>> - return find_working_add_entry_url_stuffs(
>> - svn_path_dirname(adm_access_path, scratch_pool),
>> - entry,
>> - svn_path_join(svn_path_basename(adm_access_path,
>> - scratch_pool),
>> - relative_path, scratch_pool),
>> - result_pool, scratch_pool);
>> -}
>> -
>> -
>> static svn_error_t *
>> determine_incomplete(svn_boolean_t *incomplete,
>> svn_sqlite__db_t *sdb,
>> @@ -944,7 +890,6 @@ read_entries(svn_wc_adm_access_t *adm_ac
>> apr_hash_t *working_nodes;
>> apr_hash_t *actual_nodes;
>> svn_sqlite__db_t *wc_db;
>> - apr_hash_index_t *hi;
>> apr_pool_t *result_pool;
>> apr_hash_t *entries;
>> const char *wc_db_path;
>> @@ -961,6 +906,8 @@ read_entries(svn_wc_adm_access_t *adm_ac
>>
>> result_pool = svn_wc_adm_access_pool(adm_access);
>> entries = apr_hash_make(result_pool);
>> +
>> + /* ### need database to determine: incomplete, keep_local, ACTUAL info. */
>> wc_db_path = db_path(svn_wc_adm_access_path(adm_access), scratch_pool);
>>
>> /* Open the wc.db sqlite database. */
>> @@ -969,17 +916,8 @@ read_entries(svn_wc_adm_access_t *adm_ac
>> SVN_WC__VERSION_EXPERIMENTAL, upgrade_sql,
>> scratch_pool, scratch_pool));
>>
>> - /* The basic strategy here is to get all the node information from the
>> - database for the directory in question and convert that to
>> - svn_wc_entry_t structs. To do that, we fetch each of the nodes from
>> - the three node tables into a hash, then iterate over them, linking them
>> - together as required.
>> -
>> - TODO: A smarter way would be to craft a query using the correct type of
>> - outer join so that we can get all the nodes in one fell swoop. However,
>> - that takes more thought and effort than I'm willing to invest right now.
>> - We can put it on the stack of future optimizations. */
>> -
>> + /* ### some of the data is not in the wc_db interface. grab it manually.
>> + ### trim back the columns fetched? */
>> SVN_ERR(fetch_working_nodes(&working_nodes, wc_db, scratch_pool,
>> scratch_pool));
>> SVN_ERR(fetch_actual_nodes(&actual_nodes, wc_db, scratch_pool, scratch_pool));
>> @@ -988,26 +926,11 @@ read_entries(svn_wc_adm_access_t *adm_ac
>> svn_wc_adm_access_path(adm_access),
>> scratch_pool));
>>
>> - SVN_ERR(svn_wc__db_base_get_children(&children, db,
>> - local_abspath,
>> - result_pool, scratch_pool));
>> + SVN_ERR(svn_wc__db_read_children(&children, db,
>> + local_abspath,
>> + result_pool, scratch_pool));
>>
>> - /* Is the directory also present in the BASE_NODE table? */
>> - {
>> - svn_sqlite__stmt_t *stmt;
>> - svn_boolean_t have_row;
>> -
>> - SVN_ERR(svn_sqlite__get_statement(&stmt, wc_db,
>> - STMT_SELECT_BASE_NODE_DIR_PRESENT));
>> - SVN_ERR(svn_sqlite__step(&have_row, stmt));
>> - SVN_ERR(svn_sqlite__reset(stmt));
>> -
>> - if (have_row)
>> - {
>> - /* Yup. Found it. Create an entry for this directory. */
>> - APR_ARRAY_PUSH((apr_array_header_t *)children, const char *) = "";
>> - }
>> - }
>> + APR_ARRAY_PUSH((apr_array_header_t *)children, const char *) = "";
>>
>> for (i = children->nelts; i--; )
>> {
>> @@ -1016,15 +939,19 @@ read_entries(svn_wc_adm_access_t *adm_ac
>> const char *repos_relpath;
>> svn_checksum_t *checksum;
>> svn_filesize_t translated_size;
>> - const db_working_node_t *working_node;
>> - const db_actual_node_t *actual_node;
>> svn_wc_entry_t *entry = alloc_entry(result_pool);
>> + const char *entry_abspath;
>> + const char *original_repos_relpath;
>> + const char *original_root_url;
>> + svn_boolean_t base_shadowed;
>> +
>> + svn_pool_clear(iterpool);
>>
>> entry->name = APR_ARRAY_IDX(children, i, const char *);
>>
>> - svn_pool_clear(iterpool);
>> + entry_abspath = svn_dirent_join(local_abspath, entry->name, iterpool);
>>
>> - SVN_ERR(svn_wc__db_base_get_info(
>> + SVN_ERR(svn_wc__db_read_info(
>> &status,
>> &kind,
>> &entry->revision,
>> @@ -1038,87 +965,157 @@ read_entries(svn_wc_adm_access_t *adm_ac
>> &checksum,
>> &translated_size,
>> NULL,
>> + &entry->changelist,
>> + &original_repos_relpath,
>> + &original_root_url,
>> + NULL,
>> + &entry->copyfrom_rev,
>> + NULL,
>> + NULL,
>> + &base_shadowed,
>> db,
>> - svn_dirent_join(local_abspath, entry->name, iterpool),
>> + entry_abspath,
>> result_pool,
>> iterpool));
>>
>> - /* Grab inherited repository information, if necessary. */
>> - if (repos_relpath == NULL)
>> + if (status == svn_wc__db_status_normal)
>> {
>> - SVN_ERR(svn_wc__db_scan_base_repos(&repos_relpath,
>> - &entry->repos,
>> - &entry->uuid,
>> - db,
>> - svn_dirent_join(local_abspath,
>> - entry->name,
>> - iterpool),
>> - result_pool,
>> - iterpool));
>> - }
>> + /* Plain old BASE node. */
>> + entry->schedule = svn_wc_schedule_normal;
>>
>> - /* ### most of the higher levels seem to want "infinity" for files.
>> - ### without this, it seems a report with depth=unknown was sent
>> - ### to the server, which then choked. */
>> - if (kind == svn_wc__db_kind_file)
>> - entry->depth = svn_depth_infinity;
>> + /* Grab inherited repository information, if necessary. */
>> + if (repos_relpath == NULL)
>> + {
>> + SVN_ERR(svn_wc__db_scan_base_repos(&repos_relpath,
>> + &entry->repos,
>> + &entry->uuid,
>> + db,
>> + entry_abspath,
>> + result_pool,
>> + iterpool));
>> + }
>>
>> - /* Get any corresponding working and actual nodes, removing them from
>> - their respective hashs to indicate we've seen them.
>> + /* ### hacky hacky */
>> + SVN_ERR(determine_incomplete(&entry->incomplete, wc_db,
>> + 1 /* wc_id */, entry->name));
>> + }
>> + else if (status == svn_wc__db_status_deleted)
>> + {
>> + const db_working_node_t *working_node;
>>
>> - ### these are indexed by local_relpath, which is the same as NAME */
>> - working_node = apr_hash_get(working_nodes,
>> - entry->name, APR_HASH_KEY_STRING);
>> - apr_hash_set(working_nodes, entry->name, APR_HASH_KEY_STRING, NULL);
>> - actual_node = apr_hash_get(actual_nodes,
>> - entry->name, APR_HASH_KEY_STRING);
>> - apr_hash_set(actual_nodes, entry->name, APR_HASH_KEY_STRING, NULL);
>> + /* ### we don't have to worry about moves, so this is a delete. */
>> + entry->schedule = svn_wc_schedule_delete;
>>
>> - if (working_node)
>> + /* ### keep_local */
>> + working_node = apr_hash_get(working_nodes,
>> + entry->name, APR_HASH_KEY_STRING);
>> + if (working_node && working_node->keep_local)
>> + entry->keep_local = TRUE;
>> + }
>> + else if (status == svn_wc__db_status_added)
>> {
>> - if (working_node->presence == svn_wc__db_status_not_present)
>> - entry->schedule = svn_wc_schedule_delete;
>> - else
>> + svn_wc__db_status_t work_status;
>> +
>> + if (base_shadowed)
>> entry->schedule = svn_wc_schedule_replace;
>> + else
>> + entry->schedule = svn_wc_schedule_add;
>> +
>> + SVN_ERR(svn_wc__db_scan_working(&work_status,
>> + NULL,
>> + &repos_relpath,
>> + &entry->repos,
>> + &entry->uuid,
>> + NULL, NULL, NULL, NULL,
>> + NULL,
>> + db,
>> + entry_abspath,
>> + result_pool,
>> + iterpool));
>> +
>> + if (work_status == svn_wc__db_status_copied)
>> + {
>> + entry->copied = TRUE;
>> + /* ### do children need to be schedule_normal? */
>> + }
>> + if (original_repos_relpath != NULL)
>> + {
>> + entry->copyfrom_url =
>> + svn_path_url_add_component2(original_root_url,
>> + original_repos_relpath,
>> + result_pool);
>> + }
>> +
>> + /* ### for some reason, added nodes are supposed to be rev==0. */
>> + entry->revision = 0;
>> + }
>> + else if (status == svn_wc__db_status_not_present)
>> + {
>> + entry->schedule = svn_wc_schedule_delete;
>> + entry->deleted = TRUE;
>> }
>> else
>> {
>> - entry->schedule = svn_wc_schedule_normal;
>> + /* One of the not-present varieties. Skip this node. */
>> + SVN_ERR_ASSERT(status == svn_wc__db_status_absent
>> + || status == svn_wc__db_status_excluded
>> + || status == svn_wc__db_status_incomplete);
>> + continue;
>> }
>>
>> - entry->url = svn_path_join(
>> - entry->repos,
>> - svn_path_uri_encode(repos_relpath, iterpool),
>> - result_pool);
>> -
>> - if (working_node && (working_node->copyfrom_repos_path != NULL))
>> - entry->copied = TRUE;
>> -
>> - if (working_node && working_node->keep_local)
>> - entry->keep_local = TRUE;
>> + /* ### higher levels want repos information about deleted nodes, even
>> + ### tho they are not "part of" a repository any more. */
>> + if (entry->schedule == svn_wc_schedule_delete)
>> + {
>> + svn_error_t *err;
>>
>> - if (checksum)
>> - entry->checksum = svn_checksum_to_cstring(checksum, result_pool);
>> + /* Get the information from the underlying BASE node. */
>> + err = svn_wc__db_base_get_info(NULL, &kind,
>> + &entry->revision,
>> + NULL, NULL, NULL,
>> + &entry->cmt_rev,
>> + &entry->cmt_date,
>> + &entry->cmt_author,
>> + &entry->depth,
>> + &checksum,
>> + NULL,
>> + NULL,
>> + db,
>> + entry_abspath,
>> + result_pool,
>> + iterpool);
>> + if (err)
>> + {
>> + if (err->apr_err != SVN_ERR_WC_PATH_NOT_FOUND)
>> + return err;
>>
>> - if (actual_node && (actual_node->conflict_old != NULL))
>> - {
>> - entry->conflict_old = apr_pstrdup(result_pool,
>> - actual_node->conflict_old);
>> - entry->conflict_new = apr_pstrdup(result_pool,
>> - actual_node->conflict_new);
>> - entry->conflict_wrk = apr_pstrdup(result_pool,
>> - actual_node->conflict_working);
>> + /* ### no base node? ... maybe this is a deleted child
>> + ### of a copy. what to do? */
>> + svn_error_clear(err);
>> + }
>> + else
>> + {
>> + SVN_ERR(svn_wc__db_scan_base_repos(&repos_relpath,
>> + &entry->repos,
>> + &entry->uuid,
>> + db,
>> + entry_abspath,
>> + result_pool,
>> + iterpool));
>> + }
>> }
>>
>> - if (actual_node && (actual_node->prop_reject != NULL))
>> - entry->prejfile = apr_pstrdup(result_pool, actual_node->prop_reject);
>> + /* ### our writing code (currently, erroneously) puts a 0 into the
>> + ### changed_rev column. compensate for now, rather than tweaking
>> + ### the writing. */
>> + if (entry->cmt_rev == 0)
>> + entry->cmt_rev = SVN_INVALID_REVNUM;
>>
>> - if (actual_node && actual_node->changelist != NULL)
>> - entry->changelist = apr_pstrdup(result_pool, actual_node->changelist);
>> -
>> - if (actual_node && (actual_node->tree_conflict_data != NULL))
>> - entry->tree_conflict_data = apr_pstrdup(result_pool,
>> - actual_node->tree_conflict_data);
>> + /* ### most of the higher levels seem to want "infinity" for files.
>> + ### without this, it seems a report with depth=unknown was sent
>> + ### to the server, which then choked. */
>> + if (kind == svn_wc__db_kind_file)
>> + entry->depth = svn_depth_infinity;
>>
>> if (kind == svn_wc__db_kind_dir)
>> entry->kind = svn_node_dir;
>> @@ -1129,54 +1126,48 @@ read_entries(svn_wc_adm_access_t *adm_ac
>> else
>> entry->kind = svn_node_unknown;
>>
>> - if (status == svn_wc__db_status_not_present
>> - && entry->kind == svn_node_unknown)
>> - entry->deleted = TRUE;
>> + SVN_ERR_ASSERT(repos_relpath != NULL
>> + || entry->schedule == svn_wc_schedule_delete);
>> + if (repos_relpath)
>> + entry->url = svn_path_url_add_component2(entry->repos,
>> + repos_relpath,
>> + result_pool);
>>
>> - SVN_ERR(determine_incomplete(&entry->incomplete, wc_db,
>> - 1 /* wc_id */, entry->name));
>> + if (checksum)
>> + entry->checksum = svn_checksum_to_cstring(checksum, result_pool);
>>
>> - apr_hash_set(entries, entry->name, APR_HASH_KEY_STRING, entry);
>> - }
>> + /* ### there may be an ACTUAL_NODE to grab info from. really, this
>> + ### should probably only exist for added/copied files, but it
>> + ### seems to always be needed. Just do so, for now. */
>> + if (TRUE)
>> + {
>> + const db_actual_node_t *actual_node;
>>
>> - /* Loop over any additional working nodes. */
>> - for (hi = apr_hash_first(scratch_pool, working_nodes); hi;
>> - hi = apr_hash_next(hi))
>> - {
>> - const db_working_node_t *working_node;
>> - const char *rel_path;
>> - svn_wc_entry_t *entry = alloc_entry(result_pool);
>> + actual_node = apr_hash_get(actual_nodes,
>> + entry->name, APR_HASH_KEY_STRING);
>> + if (actual_node)
>> + {
>> + if (actual_node->conflict_old != NULL)
>> + {
>> + entry->conflict_old =
>> + apr_pstrdup(result_pool, actual_node->conflict_old);
>> + entry->conflict_new =
>> + apr_pstrdup(result_pool, actual_node->conflict_new);
>> + entry->conflict_wrk =
>> + apr_pstrdup(result_pool, actual_node->conflict_working);
>> + }
>>
>> - svn_pool_clear(iterpool);
>> - apr_hash_this(hi, (const void **) &rel_path, NULL,
>> - (void **) &working_node);
>> - entry->name = apr_pstrdup(result_pool, working_node->local_relpath);
>> -
>> - /* This node is in WORKING, but not in BASE, so it must be an add. */
>> - entry->schedule = svn_wc_schedule_add;
>> -
>> - if (working_node->copyfrom_repos_path != NULL)
>> - entry->copied = TRUE;
>> -
>> - entry->keep_local = working_node->keep_local;
>> -
>> - if (working_node->checksum)
>> - entry->checksum = svn_checksum_to_cstring(working_node->checksum,
>> - result_pool);
>> -
>> - SVN_ERR(find_working_add_entry_url_stuffs(
>> - entry->name[0] == 0
>> - ? svn_path_dirname(svn_wc_adm_access_path(
>> - adm_access), iterpool)
>> - : svn_wc_adm_access_path(adm_access),
>> - entry,
>> - entry->name[0] == 0
>> - ? svn_path_basename(svn_wc_adm_access_path(
>> - adm_access), iterpool)
>> - : entry->name,
>> - result_pool, iterpool));
>> - entry->kind = working_node->kind;
>> - entry->revision = 0;
>> + if (actual_node->prop_reject != NULL)
>> + entry->prejfile =
>> + apr_pstrdup(result_pool, actual_node->prop_reject);
>> +
>> + if (actual_node->tree_conflict_data != NULL)
>> + entry->tree_conflict_data =
>> + apr_pstrdup(result_pool, actual_node->tree_conflict_data);
>> + }
>> + }
>> +
>> + /* ### do something with translated_size */
>>
>> apr_hash_set(entries, entry->name, APR_HASH_KEY_STRING, entry);
>> }
>>
>> ------------------------------------------------------
>> http://subversion.tigris.org/ds/viewMessage.do?dsForumId=495&dsMessageId=1250790
>>
>
------------------------------------------------------
http://subversion.tigris.org/ds/viewMessage.do?dsForumId=462&dsMessageId=1251106
Received on 2009-03-01 17:18:48 CET