|  | @@ -469,7 +469,10 @@ class Archive:
 | 
	
		
			
				|  |  |          self.pipeline = DownloadPipeline(self.repository, self.key)
 | 
	
		
			
				|  |  |          self.create = create
 | 
	
		
			
				|  |  |          if self.create:
 | 
	
		
			
				|  |  | -            self.items_buffer = CacheChunkBuffer(self.cache, self.key, self.stats)
 | 
	
		
			
				|  |  | +            # Use a separate statistics counter for metadata (items, archive metadata),
 | 
	
		
			
				|  |  | +            # so that archive.stats reflects only file content statistics.
 | 
	
		
			
				|  |  | +            self.meta_stats = Statistics(output_json=False, iec=iec)
 | 
	
		
			
				|  |  | +            self.items_buffer = CacheChunkBuffer(self.cache, self.key, self.meta_stats)
 | 
	
		
			
				|  |  |              if name in manifest.archives:
 | 
	
		
			
				|  |  |                  raise self.AlreadyExists(name)
 | 
	
		
			
				|  |  |              i = 0
 | 
	
	
		
			
				|  | @@ -605,7 +608,8 @@ Utilization of max. archive size: {csize_max:.0%}
 | 
	
		
			
				|  |  |      def write_checkpoint(self):
 | 
	
		
			
				|  |  |          self.save(self.checkpoint_name)
 | 
	
		
			
				|  |  |          del self.manifest.archives[self.checkpoint_name]
 | 
	
		
			
				|  |  | -        self.cache.chunk_decref(self.id, self.stats)
 | 
	
		
			
				|  |  | +        # Use meta_stats so metadata chunks do not affect archive.stats
 | 
	
		
			
				|  |  | +        self.cache.chunk_decref(self.id, self.meta_stats)
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  |      def save(self, name=None, comment=None, timestamp=None, stats=None, additional_metadata=None):
 | 
	
		
			
				|  |  |          name = name or self.name
 | 
	
	
		
			
				|  | @@ -649,7 +653,8 @@ Utilization of max. archive size: {csize_max:.0%}
 | 
	
		
			
				|  |  |          data = self.key.pack_and_authenticate_metadata(metadata.as_dict(), context=b'archive')
 | 
	
		
			
				|  |  |          self.id = self.key.id_hash(data)
 | 
	
		
			
				|  |  |          try:
 | 
	
		
			
				|  |  | -            self.cache.add_chunk(self.id, data, self.stats)
 | 
	
		
			
				|  |  | +            # Use meta_stats so metadata chunk addition does not skew archive.stats
 | 
	
		
			
				|  |  | +            self.cache.add_chunk(self.id, data, self.meta_stats)
 | 
	
		
			
				|  |  |          except IntegrityError as err:
 | 
	
		
			
				|  |  |              err_msg = str(err)
 | 
	
		
			
				|  |  |              # hack to avoid changing the RPC protocol by introducing new (more specific) exception class
 | 
	
	
		
			
				|  | @@ -687,13 +692,14 @@ Utilization of max. archive size: {csize_max:.0%}
 | 
	
		
			
				|  |  |          if have_borg12_meta and not want_unique:
 | 
	
		
			
				|  |  |              unique_csize = 0
 | 
	
		
			
				|  |  |          else:
 | 
	
		
			
				|  |  | -            def add(id):
 | 
	
		
			
				|  |  | -                entry = cache.chunks[id]
 | 
	
		
			
				|  |  | -                archive_index.add(id, 1, entry.size, entry.csize)
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  |              archive_index = ChunkIndex()
 | 
	
		
			
				|  |  |              sync = CacheSynchronizer(archive_index)
 | 
	
		
			
				|  |  | -            add(self.id)
 | 
	
		
			
				|  |  | +            # do NOT add the archive metadata chunk (self.id) here.
 | 
	
		
			
				|  |  | +            # The metadata chunk is accounted via meta_stats during creation and must not
 | 
	
		
			
				|  |  | +            # contribute to the "This archive" deduplicated size computed by borg info.
 | 
	
		
			
				|  |  | +            # See issue #9003: make info's deduplicated size match create-time stats.
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  |              # we must escape any % char in the archive name, because we use it in a format string, see #6500
 | 
	
		
			
				|  |  |              arch_name_escd = self.name.replace('%', '%%')
 | 
	
		
			
				|  |  |              pi = ProgressIndicatorPercent(total=len(self.metadata.items),
 | 
	
	
		
			
				|  | @@ -701,7 +707,8 @@ Utilization of max. archive size: {csize_max:.0%}
 | 
	
		
			
				|  |  |                                            msgid='archive.calc_stats')
 | 
	
		
			
				|  |  |              for id, chunk in zip(self.metadata.items, self.repository.get_many(self.metadata.items)):
 | 
	
		
			
				|  |  |                  pi.show(increase=1)
 | 
	
		
			
				|  |  | -                add(id)
 | 
	
		
			
				|  |  | +                # do NOT add(id) here, this is a metadata stream chunk and should not
 | 
	
		
			
				|  |  | +                # be accounted for in stats, see comment above.
 | 
	
		
			
				|  |  |                  data = self.key.decrypt(id, chunk)
 | 
	
		
			
				|  |  |                  sync.feed(data)
 | 
	
		
			
				|  |  |              unique_csize = archive_index.stats_against(cache.chunks)[3]
 | 
	
	
		
			
				|  | @@ -965,9 +972,11 @@ Utilization of max. archive size: {csize_max:.0%}
 | 
	
		
			
				|  |  |          setattr(metadata, key, value)
 | 
	
		
			
				|  |  |          data = self.key.pack_and_authenticate_metadata(metadata.as_dict(), context=b'archive')
 | 
	
		
			
				|  |  |          new_id = self.key.id_hash(data)
 | 
	
		
			
				|  |  | -        self.cache.add_chunk(new_id, data, self.stats)
 | 
	
		
			
				|  |  | +        # Use a local Statistics(), metadata updates shall not affect stats.
 | 
	
		
			
				|  |  | +        stats = Statistics(output_json=False, iec=self.iec)
 | 
	
		
			
				|  |  | +        self.cache.add_chunk(new_id, data, stats)
 | 
	
		
			
				|  |  |          self.manifest.archives[self.name] = (new_id, metadata.time)
 | 
	
		
			
				|  |  | -        self.cache.chunk_decref(self.id, self.stats)
 | 
	
		
			
				|  |  | +        self.cache.chunk_decref(self.id, stats)
 | 
	
		
			
				|  |  |          self.id = new_id
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  |      def rename(self, name):
 |