You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@subversion.apache.org by Greg Stein <gs...@gmail.com> on 2009/03/01 16:14:48 UTC

Re: svn commit: r36217 - trunk/subversion/libsvn_wc

After this change:

Summary of test results:
  819 tests PASSED
  24 tests SKIPPED
  24 tests XFAILED
  279 tests FAILED
  2 tests XPASSED

I believe that is a slight regression from before my change, but I
can't find the prior numbers. The change moves the code in the right
trajectory, but may require some further tweaking to bring the tests
back in line.

Cheers,
-g

On Sun, Mar 1, 2009 at 16:05, Greg Stein <gs...@gmail.com> wrote:
> Author: gstein
> Date: Sun Mar  1 07:05:28 2009
> New Revision: 36217
>
> Log:
> Rebuild read_entries() around the db_read_children() and db_read_info()
> interfaces. It still peeks into the database on its own, but only for some
> very limited information that isn't exposed by wc_db (yet?).
>
> * subversion/libsvn_wc/entries.c:
>  (find_working_add_entry_url_stuffs): removed. wc_db does this now.
>  (read_entries): rebuilt. primarily, the second loop over the WORKING
>    nodes is no longer performed since we now have those children in our
>    list (from read_children). there are still quite a few hacks and other
>    oddities to overcome expectations/impedance between the entries and
>    wc_db interfaces, but those will (hopefully) disappear over time.
>
> Modified:
>   trunk/subversion/libsvn_wc/entries.c
>
> Modified: trunk/subversion/libsvn_wc/entries.c
> URL: http://svn.collab.net/viewvc/svn/trunk/subversion/libsvn_wc/entries.c?pathrev=36217&r1=36216&r2=36217
> ==============================================================================
> --- trunk/subversion/libsvn_wc/entries.c        Sun Mar  1 06:48:44 2009        (r36216)
> +++ trunk/subversion/libsvn_wc/entries.c        Sun Mar  1 07:05:28 2009        (r36217)
> @@ -856,60 +856,6 @@ fetch_wc_id(apr_int64_t *wc_id, svn_sqli
>  }
>
>
> -/* This function exists for one purpose: to find the expected future url of
> -   an entry which is schedule-add.  In a centralized metadata storage
> -   situation, this is pretty easy, but in the current one-db-per-.svn scenario,
> -   we need to jump through some hoops, so here it is. */
> -static svn_error_t *
> -find_working_add_entry_url_stuffs(const char *adm_access_path,
> -                                  svn_wc_entry_t *entry,
> -                                  const char *relative_path,
> -                                  apr_pool_t *result_pool,
> -                                  apr_pool_t *scratch_pool)
> -{
> -  const char *wc_db_path = db_path(adm_access_path, scratch_pool);
> -  svn_sqlite__stmt_t *stmt;
> -  svn_boolean_t have_row;
> -  svn_sqlite__db_t *wc_db;
> -
> -  /* Open parent database. */
> -  SVN_ERR(svn_sqlite__open(&wc_db, wc_db_path, svn_sqlite__mode_readonly,
> -                           statements,
> -                           SVN_WC__VERSION_EXPERIMENTAL, upgrade_sql,
> -                           scratch_pool, scratch_pool));
> -
> -  /* Check to see if a base_node exists for the directory. */
> -  SVN_ERR(svn_sqlite__get_statement(&stmt, wc_db,
> -                                    STMT_SELECT_BASE_NODE_BY_RELPATH));
> -  SVN_ERR(svn_sqlite__bindf(stmt, "s", SVN_WC_ENTRY_THIS_DIR));
> -  SVN_ERR(svn_sqlite__step(&have_row, stmt));
> -
> -  /* If so, cat the url with the existing relative path, put that in
> -     entry->url and return. */
> -  if (have_row)
> -    {
> -      const char *base = svn_sqlite__column_text(stmt, 0, NULL);
> -
> -      entry->repos = svn_sqlite__column_text(stmt, 1, result_pool);
> -      entry->uuid = svn_sqlite__column_text(stmt, 2, result_pool);
> -      entry->url = svn_path_join_many(result_pool, entry->repos, base,
> -                                      relative_path, NULL);
> -      return svn_sqlite__reset(stmt);
> -    }
> -  SVN_ERR(svn_sqlite__reset(stmt));
> -
> -  /* If not, move a path segement from adm_access_path to relative_path and
> -     recurse. */
> -  return find_working_add_entry_url_stuffs(
> -                    svn_path_dirname(adm_access_path, scratch_pool),
> -                    entry,
> -                    svn_path_join(svn_path_basename(adm_access_path,
> -                                                    scratch_pool),
> -                                  relative_path, scratch_pool),
> -                    result_pool, scratch_pool);
> -}
> -
> -
>  static svn_error_t *
>  determine_incomplete(svn_boolean_t *incomplete,
>                      svn_sqlite__db_t *sdb,
> @@ -944,7 +890,6 @@ read_entries(svn_wc_adm_access_t *adm_ac
>   apr_hash_t *working_nodes;
>   apr_hash_t *actual_nodes;
>   svn_sqlite__db_t *wc_db;
> -  apr_hash_index_t *hi;
>   apr_pool_t *result_pool;
>   apr_hash_t *entries;
>   const char *wc_db_path;
> @@ -961,6 +906,8 @@ read_entries(svn_wc_adm_access_t *adm_ac
>
>   result_pool = svn_wc_adm_access_pool(adm_access);
>   entries = apr_hash_make(result_pool);
> +
> +  /* ### need database to determine: incomplete, keep_local, ACTUAL info.  */
>   wc_db_path = db_path(svn_wc_adm_access_path(adm_access), scratch_pool);
>
>   /* Open the wc.db sqlite database. */
> @@ -969,17 +916,8 @@ read_entries(svn_wc_adm_access_t *adm_ac
>                            SVN_WC__VERSION_EXPERIMENTAL, upgrade_sql,
>                            scratch_pool, scratch_pool));
>
> -  /* The basic strategy here is to get all the node information from the
> -     database for the directory in question and convert that to
> -     svn_wc_entry_t structs.  To do that, we fetch each of the nodes from
> -     the three node tables into a hash, then iterate over them, linking them
> -     together as required.
> -
> -     TODO: A smarter way would be to craft a query using the correct type of
> -     outer join so that we can get all the nodes in one fell swoop.  However,
> -     that takes more thought and effort than I'm willing to invest right now.
> -     We can put it on the stack of future optimizations. */
> -
> +  /* ### some of the data is not in the wc_db interface. grab it manually.
> +     ### trim back the columns fetched?  */
>   SVN_ERR(fetch_working_nodes(&working_nodes, wc_db, scratch_pool,
>                               scratch_pool));
>   SVN_ERR(fetch_actual_nodes(&actual_nodes, wc_db, scratch_pool, scratch_pool));
> @@ -988,26 +926,11 @@ read_entries(svn_wc_adm_access_t *adm_ac
>                                   svn_wc_adm_access_path(adm_access),
>                                   scratch_pool));
>
> -  SVN_ERR(svn_wc__db_base_get_children(&children, db,
> -                                       local_abspath,
> -                                       result_pool, scratch_pool));
> +  SVN_ERR(svn_wc__db_read_children(&children, db,
> +                                   local_abspath,
> +                                   result_pool, scratch_pool));
>
> -  /* Is the directory also present in the BASE_NODE table? */
> -  {
> -    svn_sqlite__stmt_t *stmt;
> -    svn_boolean_t have_row;
> -
> -    SVN_ERR(svn_sqlite__get_statement(&stmt, wc_db,
> -                                      STMT_SELECT_BASE_NODE_DIR_PRESENT));
> -    SVN_ERR(svn_sqlite__step(&have_row, stmt));
> -    SVN_ERR(svn_sqlite__reset(stmt));
> -
> -    if (have_row)
> -      {
> -        /* Yup. Found it. Create an entry for this directory. */
> -        APR_ARRAY_PUSH((apr_array_header_t *)children, const char *) = "";
> -      }
> -  }
> +  APR_ARRAY_PUSH((apr_array_header_t *)children, const char *) = "";
>
>   for (i = children->nelts; i--; )
>     {
> @@ -1016,15 +939,19 @@ read_entries(svn_wc_adm_access_t *adm_ac
>       const char *repos_relpath;
>       svn_checksum_t *checksum;
>       svn_filesize_t translated_size;
> -      const db_working_node_t *working_node;
> -      const db_actual_node_t *actual_node;
>       svn_wc_entry_t *entry = alloc_entry(result_pool);
> +      const char *entry_abspath;
> +      const char *original_repos_relpath;
> +      const char *original_root_url;
> +      svn_boolean_t base_shadowed;
> +
> +      svn_pool_clear(iterpool);
>
>       entry->name = APR_ARRAY_IDX(children, i, const char *);
>
> -      svn_pool_clear(iterpool);
> +      entry_abspath = svn_dirent_join(local_abspath, entry->name, iterpool);
>
> -      SVN_ERR(svn_wc__db_base_get_info(
> +      SVN_ERR(svn_wc__db_read_info(
>                 &status,
>                 &kind,
>                 &entry->revision,
> @@ -1038,87 +965,157 @@ read_entries(svn_wc_adm_access_t *adm_ac
>                 &checksum,
>                 &translated_size,
>                 NULL,
> +                &entry->changelist,
> +                &original_repos_relpath,
> +                &original_root_url,
> +                NULL,
> +                &entry->copyfrom_rev,
> +                NULL,
> +                NULL,
> +                &base_shadowed,
>                 db,
> -                svn_dirent_join(local_abspath, entry->name, iterpool),
> +                entry_abspath,
>                 result_pool,
>                 iterpool));
>
> -      /* Grab inherited repository information, if necessary. */
> -      if (repos_relpath == NULL)
> +      if (status == svn_wc__db_status_normal)
>         {
> -          SVN_ERR(svn_wc__db_scan_base_repos(&repos_relpath,
> -                                             &entry->repos,
> -                                             &entry->uuid,
> -                                             db,
> -                                             svn_dirent_join(local_abspath,
> -                                                             entry->name,
> -                                                             iterpool),
> -                                             result_pool,
> -                                             iterpool));
> -        }
> +          /* Plain old BASE node.  */
> +          entry->schedule = svn_wc_schedule_normal;
>
> -      /* ### most of the higher levels seem to want "infinity" for files.
> -         ### without this, it seems a report with depth=unknown was sent
> -         ### to the server, which then choked.  */
> -      if (kind == svn_wc__db_kind_file)
> -        entry->depth = svn_depth_infinity;
> +          /* Grab inherited repository information, if necessary. */
> +          if (repos_relpath == NULL)
> +            {
> +              SVN_ERR(svn_wc__db_scan_base_repos(&repos_relpath,
> +                                                 &entry->repos,
> +                                                 &entry->uuid,
> +                                                 db,
> +                                                 entry_abspath,
> +                                                 result_pool,
> +                                                 iterpool));
> +            }
>
> -      /* Get any corresponding working and actual nodes, removing them from
> -         their respective hashs to indicate we've seen them.
> +          /* ### hacky hacky  */
> +          SVN_ERR(determine_incomplete(&entry->incomplete, wc_db,
> +                                       1 /* wc_id */, entry->name));
> +        }
> +      else if (status == svn_wc__db_status_deleted)
> +        {
> +          const db_working_node_t *working_node;
>
> -         ### these are indexed by local_relpath, which is the same as NAME  */
> -      working_node = apr_hash_get(working_nodes,
> -                                  entry->name, APR_HASH_KEY_STRING);
> -      apr_hash_set(working_nodes, entry->name, APR_HASH_KEY_STRING, NULL);
> -      actual_node = apr_hash_get(actual_nodes,
> -                                 entry->name, APR_HASH_KEY_STRING);
> -      apr_hash_set(actual_nodes, entry->name, APR_HASH_KEY_STRING, NULL);
> +          /* ### we don't have to worry about moves, so this is a delete. */
> +          entry->schedule = svn_wc_schedule_delete;
>
> -      if (working_node)
> +          /* ### keep_local */
> +          working_node = apr_hash_get(working_nodes,
> +                                      entry->name, APR_HASH_KEY_STRING);
> +          if (working_node && working_node->keep_local)
> +            entry->keep_local = TRUE;
> +        }
> +      else if (status == svn_wc__db_status_added)
>         {
> -          if (working_node->presence == svn_wc__db_status_not_present)
> -            entry->schedule = svn_wc_schedule_delete;
> -          else
> +          svn_wc__db_status_t work_status;
> +
> +          if (base_shadowed)
>             entry->schedule = svn_wc_schedule_replace;
> +          else
> +            entry->schedule = svn_wc_schedule_add;
> +
> +          SVN_ERR(svn_wc__db_scan_working(&work_status,
> +                                          NULL,
> +                                          &repos_relpath,
> +                                          &entry->repos,
> +                                          &entry->uuid,
> +                                          NULL, NULL, NULL, NULL,
> +                                          NULL,
> +                                          db,
> +                                          entry_abspath,
> +                                          result_pool,
> +                                          iterpool));
> +
> +          if (work_status == svn_wc__db_status_copied)
> +            {
> +              entry->copied = TRUE;
> +              /* ### do children need to be schedule_normal?  */
> +            }
> +          if (original_repos_relpath != NULL)
> +            {
> +              entry->copyfrom_url =
> +                svn_path_url_add_component2(original_root_url,
> +                                            original_repos_relpath,
> +                                            result_pool);
> +            }
> +
> +          /* ### for some reason, added nodes are supposed to be rev==0.  */
> +          entry->revision = 0;
> +        }
> +      else if (status == svn_wc__db_status_not_present)
> +        {
> +          entry->schedule = svn_wc_schedule_delete;
> +          entry->deleted = TRUE;
>         }
>       else
>         {
> -          entry->schedule = svn_wc_schedule_normal;
> +          /* One of the not-present varieties. Skip this node.  */
> +          SVN_ERR_ASSERT(status == svn_wc__db_status_absent
> +                         || status == svn_wc__db_status_excluded
> +                         || status == svn_wc__db_status_incomplete);
> +          continue;
>         }
>
> -      entry->url = svn_path_join(
> -                    entry->repos,
> -                    svn_path_uri_encode(repos_relpath, iterpool),
> -                    result_pool);
> -
> -      if (working_node && (working_node->copyfrom_repos_path != NULL))
> -        entry->copied = TRUE;
> -
> -      if (working_node && working_node->keep_local)
> -        entry->keep_local = TRUE;
> +      /* ### higher levels want repos information about deleted nodes, even
> +         ### tho they are not "part of" a repository any more.  */
> +      if (entry->schedule == svn_wc_schedule_delete)
> +        {
> +          svn_error_t *err;
>
> -      if (checksum)
> -        entry->checksum = svn_checksum_to_cstring(checksum, result_pool);
> +          /* Get the information from the underlying BASE node.  */
> +          err = svn_wc__db_base_get_info(NULL, &kind,
> +                                         &entry->revision,
> +                                         NULL, NULL, NULL,
> +                                         &entry->cmt_rev,
> +                                         &entry->cmt_date,
> +                                         &entry->cmt_author,
> +                                         &entry->depth,
> +                                         &checksum,
> +                                         NULL,
> +                                         NULL,
> +                                         db,
> +                                         entry_abspath,
> +                                         result_pool,
> +                                         iterpool);
> +          if (err)
> +            {
> +              if (err->apr_err != SVN_ERR_WC_PATH_NOT_FOUND)
> +                return err;
>
> -      if (actual_node && (actual_node->conflict_old != NULL))
> -        {
> -          entry->conflict_old = apr_pstrdup(result_pool,
> -                                            actual_node->conflict_old);
> -          entry->conflict_new = apr_pstrdup(result_pool,
> -                                            actual_node->conflict_new);
> -          entry->conflict_wrk = apr_pstrdup(result_pool,
> -                                            actual_node->conflict_working);
> +              /* ### no base node? ... maybe this is a deleted child
> +                 ### of a copy.  what to do?  */
> +              svn_error_clear(err);
> +            }
> +          else
> +            {
> +              SVN_ERR(svn_wc__db_scan_base_repos(&repos_relpath,
> +                                                 &entry->repos,
> +                                                 &entry->uuid,
> +                                                 db,
> +                                                 entry_abspath,
> +                                                 result_pool,
> +                                                 iterpool));
> +            }
>         }
>
> -      if (actual_node && (actual_node->prop_reject != NULL))
> -        entry->prejfile = apr_pstrdup(result_pool, actual_node->prop_reject);
> +      /* ### our writing code (currently, erroneously) puts a 0 into the
> +         ### changed_rev column. compensate for now, rather than tweaking
> +         ### the writing.  */
> +      if (entry->cmt_rev == 0)
> +        entry->cmt_rev = SVN_INVALID_REVNUM;
>
> -      if (actual_node && actual_node->changelist != NULL)
> -        entry->changelist = apr_pstrdup(result_pool, actual_node->changelist);
> -
> -      if (actual_node && (actual_node->tree_conflict_data != NULL))
> -        entry->tree_conflict_data = apr_pstrdup(result_pool,
> -                                              actual_node->tree_conflict_data);
> +      /* ### most of the higher levels seem to want "infinity" for files.
> +         ### without this, it seems a report with depth=unknown was sent
> +         ### to the server, which then choked.  */
> +      if (kind == svn_wc__db_kind_file)
> +        entry->depth = svn_depth_infinity;
>
>       if (kind == svn_wc__db_kind_dir)
>         entry->kind = svn_node_dir;
> @@ -1129,54 +1126,48 @@ read_entries(svn_wc_adm_access_t *adm_ac
>       else
>         entry->kind = svn_node_unknown;
>
> -      if (status == svn_wc__db_status_not_present
> -            && entry->kind == svn_node_unknown)
> -        entry->deleted = TRUE;
> +      SVN_ERR_ASSERT(repos_relpath != NULL
> +                     || entry->schedule == svn_wc_schedule_delete);
> +      if (repos_relpath)
> +        entry->url = svn_path_url_add_component2(entry->repos,
> +                                                 repos_relpath,
> +                                                 result_pool);
>
> -      SVN_ERR(determine_incomplete(&entry->incomplete, wc_db,
> -                                   1 /* wc_id */, entry->name));
> +      if (checksum)
> +        entry->checksum = svn_checksum_to_cstring(checksum, result_pool);
>
> -      apr_hash_set(entries, entry->name, APR_HASH_KEY_STRING, entry);
> -    }
> +      /* ### there may be an ACTUAL_NODE to grab info from.  really, this
> +         ### should probably only exist for added/copied files, but it
> +         ### seems to always be needed. Just do so, for now.  */
> +      if (TRUE)
> +        {
> +          const db_actual_node_t *actual_node;
>
> -  /* Loop over any additional working nodes. */
> -  for (hi = apr_hash_first(scratch_pool, working_nodes); hi;
> -        hi = apr_hash_next(hi))
> -    {
> -      const db_working_node_t *working_node;
> -      const char *rel_path;
> -      svn_wc_entry_t *entry = alloc_entry(result_pool);
> +          actual_node = apr_hash_get(actual_nodes,
> +                                     entry->name, APR_HASH_KEY_STRING);
> +          if (actual_node)
> +            {
> +              if (actual_node->conflict_old != NULL)
> +                {
> +                  entry->conflict_old =
> +                    apr_pstrdup(result_pool, actual_node->conflict_old);
> +                  entry->conflict_new =
> +                    apr_pstrdup(result_pool, actual_node->conflict_new);
> +                  entry->conflict_wrk =
> +                    apr_pstrdup(result_pool, actual_node->conflict_working);
> +                }
>
> -      svn_pool_clear(iterpool);
> -      apr_hash_this(hi, (const void **) &rel_path, NULL,
> -                    (void **) &working_node);
> -      entry->name = apr_pstrdup(result_pool, working_node->local_relpath);
> -
> -      /* This node is in WORKING, but not in BASE, so it must be an add. */
> -      entry->schedule = svn_wc_schedule_add;
> -
> -      if (working_node->copyfrom_repos_path != NULL)
> -        entry->copied = TRUE;
> -
> -      entry->keep_local = working_node->keep_local;
> -
> -      if (working_node->checksum)
> -        entry->checksum = svn_checksum_to_cstring(working_node->checksum,
> -                                                  result_pool);
> -
> -      SVN_ERR(find_working_add_entry_url_stuffs(
> -                        entry->name[0] == 0
> -                            ? svn_path_dirname(svn_wc_adm_access_path(
> -                                            adm_access), iterpool)
> -                            : svn_wc_adm_access_path(adm_access),
> -                        entry,
> -                        entry->name[0] == 0
> -                            ? svn_path_basename(svn_wc_adm_access_path(
> -                                                   adm_access), iterpool)
> -                            : entry->name,
> -                        result_pool, iterpool));
> -      entry->kind = working_node->kind;
> -      entry->revision = 0;
> +              if (actual_node->prop_reject != NULL)
> +                entry->prejfile =
> +                  apr_pstrdup(result_pool, actual_node->prop_reject);
> +
> +              if (actual_node->tree_conflict_data != NULL)
> +                entry->tree_conflict_data =
> +                  apr_pstrdup(result_pool, actual_node->tree_conflict_data);
> +            }
> +        }
> +
> +      /* ### do something with translated_size  */
>
>       apr_hash_set(entries, entry->name, APR_HASH_KEY_STRING, entry);
>     }
>
> ------------------------------------------------------
> http://subversion.tigris.org/ds/viewMessage.do?dsForumId=495&dsMessageId=1250790
>

------------------------------------------------------
http://subversion.tigris.org/ds/viewMessage.do?dsForumId=462&dsMessageId=1251097


Re: svn commit: r36217 - trunk/subversion/libsvn_wc

Posted by Greg Stein <gs...@gmail.com>.
Found it. Hyrum said 207 tests were failing a few days ago. Ah well...

Time to fix things.

On Sun, Mar 1, 2009 at 17:14, Greg Stein <gs...@gmail.com> wrote:
> After this change:
>
> Summary of test results:
>  819 tests PASSED
>  24 tests SKIPPED
>  24 tests XFAILED
>  279 tests FAILED
>  2 tests XPASSED
>
> I believe that is a slight regression from before my change, but I
> can't find the prior numbers. The change moves the code in the right
> trajectory, but may require some further tweaking to bring the tests
> back in line.
>
> Cheers,
> -g
>
> On Sun, Mar 1, 2009 at 16:05, Greg Stein <gs...@gmail.com> wrote:
>> Author: gstein
>> Date: Sun Mar  1 07:05:28 2009
>> New Revision: 36217
>>
>> Log:
>> Rebuild read_entries() around the db_read_children() and db_read_info()
>> interfaces. It still peeks into the database on its own, but only for some
>> very limited information that isn't exposed by wc_db (yet?).
>>
>> * subversion/libsvn_wc/entries.c:
>>  (find_working_add_entry_url_stuffs): removed. wc_db does this now.
>>  (read_entries): rebuilt. primarily, the second loop over the WORKING
>>    nodes is no longer performed since we now have those children in our
>>    list (from read_children). there are still quite a few hacks and other
>>    oddities to overcome expectations/impedance between the entries and
>>    wc_db interfaces, but those will (hopefully) disappear over time.
>>
>> Modified:
>>   trunk/subversion/libsvn_wc/entries.c
>>
>> Modified: trunk/subversion/libsvn_wc/entries.c
>> URL: http://svn.collab.net/viewvc/svn/trunk/subversion/libsvn_wc/entries.c?pathrev=36217&r1=36216&r2=36217
>> ==============================================================================
>> --- trunk/subversion/libsvn_wc/entries.c        Sun Mar  1 06:48:44 2009        (r36216)
>> +++ trunk/subversion/libsvn_wc/entries.c        Sun Mar  1 07:05:28 2009        (r36217)
>> @@ -856,60 +856,6 @@ fetch_wc_id(apr_int64_t *wc_id, svn_sqli
>>  }
>>
>>
>> -/* This function exists for one purpose: to find the expected future url of
>> -   an entry which is schedule-add.  In a centralized metadata storage
>> -   situation, this is pretty easy, but in the current one-db-per-.svn scenario,
>> -   we need to jump through some hoops, so here it is. */
>> -static svn_error_t *
>> -find_working_add_entry_url_stuffs(const char *adm_access_path,
>> -                                  svn_wc_entry_t *entry,
>> -                                  const char *relative_path,
>> -                                  apr_pool_t *result_pool,
>> -                                  apr_pool_t *scratch_pool)
>> -{
>> -  const char *wc_db_path = db_path(adm_access_path, scratch_pool);
>> -  svn_sqlite__stmt_t *stmt;
>> -  svn_boolean_t have_row;
>> -  svn_sqlite__db_t *wc_db;
>> -
>> -  /* Open parent database. */
>> -  SVN_ERR(svn_sqlite__open(&wc_db, wc_db_path, svn_sqlite__mode_readonly,
>> -                           statements,
>> -                           SVN_WC__VERSION_EXPERIMENTAL, upgrade_sql,
>> -                           scratch_pool, scratch_pool));
>> -
>> -  /* Check to see if a base_node exists for the directory. */
>> -  SVN_ERR(svn_sqlite__get_statement(&stmt, wc_db,
>> -                                    STMT_SELECT_BASE_NODE_BY_RELPATH));
>> -  SVN_ERR(svn_sqlite__bindf(stmt, "s", SVN_WC_ENTRY_THIS_DIR));
>> -  SVN_ERR(svn_sqlite__step(&have_row, stmt));
>> -
>> -  /* If so, cat the url with the existing relative path, put that in
>> -     entry->url and return. */
>> -  if (have_row)
>> -    {
>> -      const char *base = svn_sqlite__column_text(stmt, 0, NULL);
>> -
>> -      entry->repos = svn_sqlite__column_text(stmt, 1, result_pool);
>> -      entry->uuid = svn_sqlite__column_text(stmt, 2, result_pool);
>> -      entry->url = svn_path_join_many(result_pool, entry->repos, base,
>> -                                      relative_path, NULL);
>> -      return svn_sqlite__reset(stmt);
>> -    }
>> -  SVN_ERR(svn_sqlite__reset(stmt));
>> -
>> -  /* If not, move a path segement from adm_access_path to relative_path and
>> -     recurse. */
>> -  return find_working_add_entry_url_stuffs(
>> -                    svn_path_dirname(adm_access_path, scratch_pool),
>> -                    entry,
>> -                    svn_path_join(svn_path_basename(adm_access_path,
>> -                                                    scratch_pool),
>> -                                  relative_path, scratch_pool),
>> -                    result_pool, scratch_pool);
>> -}
>> -
>> -
>>  static svn_error_t *
>>  determine_incomplete(svn_boolean_t *incomplete,
>>                      svn_sqlite__db_t *sdb,
>> @@ -944,7 +890,6 @@ read_entries(svn_wc_adm_access_t *adm_ac
>>   apr_hash_t *working_nodes;
>>   apr_hash_t *actual_nodes;
>>   svn_sqlite__db_t *wc_db;
>> -  apr_hash_index_t *hi;
>>   apr_pool_t *result_pool;
>>   apr_hash_t *entries;
>>   const char *wc_db_path;
>> @@ -961,6 +906,8 @@ read_entries(svn_wc_adm_access_t *adm_ac
>>
>>   result_pool = svn_wc_adm_access_pool(adm_access);
>>   entries = apr_hash_make(result_pool);
>> +
>> +  /* ### need database to determine: incomplete, keep_local, ACTUAL info.  */
>>   wc_db_path = db_path(svn_wc_adm_access_path(adm_access), scratch_pool);
>>
>>   /* Open the wc.db sqlite database. */
>> @@ -969,17 +916,8 @@ read_entries(svn_wc_adm_access_t *adm_ac
>>                            SVN_WC__VERSION_EXPERIMENTAL, upgrade_sql,
>>                            scratch_pool, scratch_pool));
>>
>> -  /* The basic strategy here is to get all the node information from the
>> -     database for the directory in question and convert that to
>> -     svn_wc_entry_t structs.  To do that, we fetch each of the nodes from
>> -     the three node tables into a hash, then iterate over them, linking them
>> -     together as required.
>> -
>> -     TODO: A smarter way would be to craft a query using the correct type of
>> -     outer join so that we can get all the nodes in one fell swoop.  However,
>> -     that takes more thought and effort than I'm willing to invest right now.
>> -     We can put it on the stack of future optimizations. */
>> -
>> +  /* ### some of the data is not in the wc_db interface. grab it manually.
>> +     ### trim back the columns fetched?  */
>>   SVN_ERR(fetch_working_nodes(&working_nodes, wc_db, scratch_pool,
>>                               scratch_pool));
>>   SVN_ERR(fetch_actual_nodes(&actual_nodes, wc_db, scratch_pool, scratch_pool));
>> @@ -988,26 +926,11 @@ read_entries(svn_wc_adm_access_t *adm_ac
>>                                   svn_wc_adm_access_path(adm_access),
>>                                   scratch_pool));
>>
>> -  SVN_ERR(svn_wc__db_base_get_children(&children, db,
>> -                                       local_abspath,
>> -                                       result_pool, scratch_pool));
>> +  SVN_ERR(svn_wc__db_read_children(&children, db,
>> +                                   local_abspath,
>> +                                   result_pool, scratch_pool));
>>
>> -  /* Is the directory also present in the BASE_NODE table? */
>> -  {
>> -    svn_sqlite__stmt_t *stmt;
>> -    svn_boolean_t have_row;
>> -
>> -    SVN_ERR(svn_sqlite__get_statement(&stmt, wc_db,
>> -                                      STMT_SELECT_BASE_NODE_DIR_PRESENT));
>> -    SVN_ERR(svn_sqlite__step(&have_row, stmt));
>> -    SVN_ERR(svn_sqlite__reset(stmt));
>> -
>> -    if (have_row)
>> -      {
>> -        /* Yup. Found it. Create an entry for this directory. */
>> -        APR_ARRAY_PUSH((apr_array_header_t *)children, const char *) = "";
>> -      }
>> -  }
>> +  APR_ARRAY_PUSH((apr_array_header_t *)children, const char *) = "";
>>
>>   for (i = children->nelts; i--; )
>>     {
>> @@ -1016,15 +939,19 @@ read_entries(svn_wc_adm_access_t *adm_ac
>>       const char *repos_relpath;
>>       svn_checksum_t *checksum;
>>       svn_filesize_t translated_size;
>> -      const db_working_node_t *working_node;
>> -      const db_actual_node_t *actual_node;
>>       svn_wc_entry_t *entry = alloc_entry(result_pool);
>> +      const char *entry_abspath;
>> +      const char *original_repos_relpath;
>> +      const char *original_root_url;
>> +      svn_boolean_t base_shadowed;
>> +
>> +      svn_pool_clear(iterpool);
>>
>>       entry->name = APR_ARRAY_IDX(children, i, const char *);
>>
>> -      svn_pool_clear(iterpool);
>> +      entry_abspath = svn_dirent_join(local_abspath, entry->name, iterpool);
>>
>> -      SVN_ERR(svn_wc__db_base_get_info(
>> +      SVN_ERR(svn_wc__db_read_info(
>>                 &status,
>>                 &kind,
>>                 &entry->revision,
>> @@ -1038,87 +965,157 @@ read_entries(svn_wc_adm_access_t *adm_ac
>>                 &checksum,
>>                 &translated_size,
>>                 NULL,
>> +                &entry->changelist,
>> +                &original_repos_relpath,
>> +                &original_root_url,
>> +                NULL,
>> +                &entry->copyfrom_rev,
>> +                NULL,
>> +                NULL,
>> +                &base_shadowed,
>>                 db,
>> -                svn_dirent_join(local_abspath, entry->name, iterpool),
>> +                entry_abspath,
>>                 result_pool,
>>                 iterpool));
>>
>> -      /* Grab inherited repository information, if necessary. */
>> -      if (repos_relpath == NULL)
>> +      if (status == svn_wc__db_status_normal)
>>         {
>> -          SVN_ERR(svn_wc__db_scan_base_repos(&repos_relpath,
>> -                                             &entry->repos,
>> -                                             &entry->uuid,
>> -                                             db,
>> -                                             svn_dirent_join(local_abspath,
>> -                                                             entry->name,
>> -                                                             iterpool),
>> -                                             result_pool,
>> -                                             iterpool));
>> -        }
>> +          /* Plain old BASE node.  */
>> +          entry->schedule = svn_wc_schedule_normal;
>>
>> -      /* ### most of the higher levels seem to want "infinity" for files.
>> -         ### without this, it seems a report with depth=unknown was sent
>> -         ### to the server, which then choked.  */
>> -      if (kind == svn_wc__db_kind_file)
>> -        entry->depth = svn_depth_infinity;
>> +          /* Grab inherited repository information, if necessary. */
>> +          if (repos_relpath == NULL)
>> +            {
>> +              SVN_ERR(svn_wc__db_scan_base_repos(&repos_relpath,
>> +                                                 &entry->repos,
>> +                                                 &entry->uuid,
>> +                                                 db,
>> +                                                 entry_abspath,
>> +                                                 result_pool,
>> +                                                 iterpool));
>> +            }
>>
>> -      /* Get any corresponding working and actual nodes, removing them from
>> -         their respective hashs to indicate we've seen them.
>> +          /* ### hacky hacky  */
>> +          SVN_ERR(determine_incomplete(&entry->incomplete, wc_db,
>> +                                       1 /* wc_id */, entry->name));
>> +        }
>> +      else if (status == svn_wc__db_status_deleted)
>> +        {
>> +          const db_working_node_t *working_node;
>>
>> -         ### these are indexed by local_relpath, which is the same as NAME  */
>> -      working_node = apr_hash_get(working_nodes,
>> -                                  entry->name, APR_HASH_KEY_STRING);
>> -      apr_hash_set(working_nodes, entry->name, APR_HASH_KEY_STRING, NULL);
>> -      actual_node = apr_hash_get(actual_nodes,
>> -                                 entry->name, APR_HASH_KEY_STRING);
>> -      apr_hash_set(actual_nodes, entry->name, APR_HASH_KEY_STRING, NULL);
>> +          /* ### we don't have to worry about moves, so this is a delete. */
>> +          entry->schedule = svn_wc_schedule_delete;
>>
>> -      if (working_node)
>> +          /* ### keep_local */
>> +          working_node = apr_hash_get(working_nodes,
>> +                                      entry->name, APR_HASH_KEY_STRING);
>> +          if (working_node && working_node->keep_local)
>> +            entry->keep_local = TRUE;
>> +        }
>> +      else if (status == svn_wc__db_status_added)
>>         {
>> -          if (working_node->presence == svn_wc__db_status_not_present)
>> -            entry->schedule = svn_wc_schedule_delete;
>> -          else
>> +          svn_wc__db_status_t work_status;
>> +
>> +          if (base_shadowed)
>>             entry->schedule = svn_wc_schedule_replace;
>> +          else
>> +            entry->schedule = svn_wc_schedule_add;
>> +
>> +          SVN_ERR(svn_wc__db_scan_working(&work_status,
>> +                                          NULL,
>> +                                          &repos_relpath,
>> +                                          &entry->repos,
>> +                                          &entry->uuid,
>> +                                          NULL, NULL, NULL, NULL,
>> +                                          NULL,
>> +                                          db,
>> +                                          entry_abspath,
>> +                                          result_pool,
>> +                                          iterpool));
>> +
>> +          if (work_status == svn_wc__db_status_copied)
>> +            {
>> +              entry->copied = TRUE;
>> +              /* ### do children need to be schedule_normal?  */
>> +            }
>> +          if (original_repos_relpath != NULL)
>> +            {
>> +              entry->copyfrom_url =
>> +                svn_path_url_add_component2(original_root_url,
>> +                                            original_repos_relpath,
>> +                                            result_pool);
>> +            }
>> +
>> +          /* ### for some reason, added nodes are supposed to be rev==0.  */
>> +          entry->revision = 0;
>> +        }
>> +      else if (status == svn_wc__db_status_not_present)
>> +        {
>> +          entry->schedule = svn_wc_schedule_delete;
>> +          entry->deleted = TRUE;
>>         }
>>       else
>>         {
>> -          entry->schedule = svn_wc_schedule_normal;
>> +          /* One of the not-present varieties. Skip this node.  */
>> +          SVN_ERR_ASSERT(status == svn_wc__db_status_absent
>> +                         || status == svn_wc__db_status_excluded
>> +                         || status == svn_wc__db_status_incomplete);
>> +          continue;
>>         }
>>
>> -      entry->url = svn_path_join(
>> -                    entry->repos,
>> -                    svn_path_uri_encode(repos_relpath, iterpool),
>> -                    result_pool);
>> -
>> -      if (working_node && (working_node->copyfrom_repos_path != NULL))
>> -        entry->copied = TRUE;
>> -
>> -      if (working_node && working_node->keep_local)
>> -        entry->keep_local = TRUE;
>> +      /* ### higher levels want repos information about deleted nodes, even
>> +         ### tho they are not "part of" a repository any more.  */
>> +      if (entry->schedule == svn_wc_schedule_delete)
>> +        {
>> +          svn_error_t *err;
>>
>> -      if (checksum)
>> -        entry->checksum = svn_checksum_to_cstring(checksum, result_pool);
>> +          /* Get the information from the underlying BASE node.  */
>> +          err = svn_wc__db_base_get_info(NULL, &kind,
>> +                                         &entry->revision,
>> +                                         NULL, NULL, NULL,
>> +                                         &entry->cmt_rev,
>> +                                         &entry->cmt_date,
>> +                                         &entry->cmt_author,
>> +                                         &entry->depth,
>> +                                         &checksum,
>> +                                         NULL,
>> +                                         NULL,
>> +                                         db,
>> +                                         entry_abspath,
>> +                                         result_pool,
>> +                                         iterpool);
>> +          if (err)
>> +            {
>> +              if (err->apr_err != SVN_ERR_WC_PATH_NOT_FOUND)
>> +                return err;
>>
>> -      if (actual_node && (actual_node->conflict_old != NULL))
>> -        {
>> -          entry->conflict_old = apr_pstrdup(result_pool,
>> -                                            actual_node->conflict_old);
>> -          entry->conflict_new = apr_pstrdup(result_pool,
>> -                                            actual_node->conflict_new);
>> -          entry->conflict_wrk = apr_pstrdup(result_pool,
>> -                                            actual_node->conflict_working);
>> +              /* ### no base node? ... maybe this is a deleted child
>> +                 ### of a copy.  what to do?  */
>> +              svn_error_clear(err);
>> +            }
>> +          else
>> +            {
>> +              SVN_ERR(svn_wc__db_scan_base_repos(&repos_relpath,
>> +                                                 &entry->repos,
>> +                                                 &entry->uuid,
>> +                                                 db,
>> +                                                 entry_abspath,
>> +                                                 result_pool,
>> +                                                 iterpool));
>> +            }
>>         }
>>
>> -      if (actual_node && (actual_node->prop_reject != NULL))
>> -        entry->prejfile = apr_pstrdup(result_pool, actual_node->prop_reject);
>> +      /* ### our writing code (currently, erroneously) puts a 0 into the
>> +         ### changed_rev column. compensate for now, rather than tweaking
>> +         ### the writing.  */
>> +      if (entry->cmt_rev == 0)
>> +        entry->cmt_rev = SVN_INVALID_REVNUM;
>>
>> -      if (actual_node && actual_node->changelist != NULL)
>> -        entry->changelist = apr_pstrdup(result_pool, actual_node->changelist);
>> -
>> -      if (actual_node && (actual_node->tree_conflict_data != NULL))
>> -        entry->tree_conflict_data = apr_pstrdup(result_pool,
>> -                                              actual_node->tree_conflict_data);
>> +      /* ### most of the higher levels seem to want "infinity" for files.
>> +         ### without this, it seems a report with depth=unknown was sent
>> +         ### to the server, which then choked.  */
>> +      if (kind == svn_wc__db_kind_file)
>> +        entry->depth = svn_depth_infinity;
>>
>>       if (kind == svn_wc__db_kind_dir)
>>         entry->kind = svn_node_dir;
>> @@ -1129,54 +1126,48 @@ read_entries(svn_wc_adm_access_t *adm_ac
>>       else
>>         entry->kind = svn_node_unknown;
>>
>> -      if (status == svn_wc__db_status_not_present
>> -            && entry->kind == svn_node_unknown)
>> -        entry->deleted = TRUE;
>> +      SVN_ERR_ASSERT(repos_relpath != NULL
>> +                     || entry->schedule == svn_wc_schedule_delete);
>> +      if (repos_relpath)
>> +        entry->url = svn_path_url_add_component2(entry->repos,
>> +                                                 repos_relpath,
>> +                                                 result_pool);
>>
>> -      SVN_ERR(determine_incomplete(&entry->incomplete, wc_db,
>> -                                   1 /* wc_id */, entry->name));
>> +      if (checksum)
>> +        entry->checksum = svn_checksum_to_cstring(checksum, result_pool);
>>
>> -      apr_hash_set(entries, entry->name, APR_HASH_KEY_STRING, entry);
>> -    }
>> +      /* ### there may be an ACTUAL_NODE to grab info from.  really, this
>> +         ### should probably only exist for added/copied files, but it
>> +         ### seems to always be needed. Just do so, for now.  */
>> +      if (TRUE)
>> +        {
>> +          const db_actual_node_t *actual_node;
>>
>> -  /* Loop over any additional working nodes. */
>> -  for (hi = apr_hash_first(scratch_pool, working_nodes); hi;
>> -        hi = apr_hash_next(hi))
>> -    {
>> -      const db_working_node_t *working_node;
>> -      const char *rel_path;
>> -      svn_wc_entry_t *entry = alloc_entry(result_pool);
>> +          actual_node = apr_hash_get(actual_nodes,
>> +                                     entry->name, APR_HASH_KEY_STRING);
>> +          if (actual_node)
>> +            {
>> +              if (actual_node->conflict_old != NULL)
>> +                {
>> +                  entry->conflict_old =
>> +                    apr_pstrdup(result_pool, actual_node->conflict_old);
>> +                  entry->conflict_new =
>> +                    apr_pstrdup(result_pool, actual_node->conflict_new);
>> +                  entry->conflict_wrk =
>> +                    apr_pstrdup(result_pool, actual_node->conflict_working);
>> +                }
>>
>> -      svn_pool_clear(iterpool);
>> -      apr_hash_this(hi, (const void **) &rel_path, NULL,
>> -                    (void **) &working_node);
>> -      entry->name = apr_pstrdup(result_pool, working_node->local_relpath);
>> -
>> -      /* This node is in WORKING, but not in BASE, so it must be an add. */
>> -      entry->schedule = svn_wc_schedule_add;
>> -
>> -      if (working_node->copyfrom_repos_path != NULL)
>> -        entry->copied = TRUE;
>> -
>> -      entry->keep_local = working_node->keep_local;
>> -
>> -      if (working_node->checksum)
>> -        entry->checksum = svn_checksum_to_cstring(working_node->checksum,
>> -                                                  result_pool);
>> -
>> -      SVN_ERR(find_working_add_entry_url_stuffs(
>> -                        entry->name[0] == 0
>> -                            ? svn_path_dirname(svn_wc_adm_access_path(
>> -                                            adm_access), iterpool)
>> -                            : svn_wc_adm_access_path(adm_access),
>> -                        entry,
>> -                        entry->name[0] == 0
>> -                            ? svn_path_basename(svn_wc_adm_access_path(
>> -                                                   adm_access), iterpool)
>> -                            : entry->name,
>> -                        result_pool, iterpool));
>> -      entry->kind = working_node->kind;
>> -      entry->revision = 0;
>> +              if (actual_node->prop_reject != NULL)
>> +                entry->prejfile =
>> +                  apr_pstrdup(result_pool, actual_node->prop_reject);
>> +
>> +              if (actual_node->tree_conflict_data != NULL)
>> +                entry->tree_conflict_data =
>> +                  apr_pstrdup(result_pool, actual_node->tree_conflict_data);
>> +            }
>> +        }
>> +
>> +      /* ### do something with translated_size  */
>>
>>       apr_hash_set(entries, entry->name, APR_HASH_KEY_STRING, entry);
>>     }
>>
>> ------------------------------------------------------
>> http://subversion.tigris.org/ds/viewMessage.do?dsForumId=495&dsMessageId=1250790
>>
>

------------------------------------------------------
http://subversion.tigris.org/ds/viewMessage.do?dsForumId=462&dsMessageId=1251106