mirror of
https://github.com/mainflux/mainflux.git
synced 2025-04-26 13:48:53 +08:00

* Return Auth service Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Update Compose to run with SpiceDB and Auth svc Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Update auth gRPC API Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Remove Users' policies Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Move Groups to internal Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Use shared groups in Users Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Remove unused code Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Use pkg Groups in Things Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Remove Things groups Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Make imports consistent Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Update Groups networking Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Remove things groups-specific API Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Move Things Clients to the root Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Move Clients to Users root Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Temporarily remove tracing Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Fix imports Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Add buffer config for gRPC Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Update auth type for Things Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Use Auth for login Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Add temporary solution for refresh token Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Update Tokenizer interface Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Updade tokens issuing Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Fix token issuing Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Update JWT validator and refactor Tokenizer Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Rename access timeout Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Rename login to authenticate Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Update Identify to use SubjectID Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Add Auth to Groups Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Use the Auth service for Groups Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Update auth schema Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Fix Auth for Groups Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Add auth for addons (#14) Signed-off-by: Arvindh <arvindh91@gmail.com> Speparate Login and Refresh tokens Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Merge authN and authZ requests for things Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Add connect and disconnect Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Update sharing Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Fix policies addition and removal Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Update relation with roels Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Add gRPC to Things Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Assign and Unassign members to group and Listing of Group members (#15) * add auth for addons Signed-off-by: Arvindh <arvindh91@gmail.com> * add assign and unassign to group Signed-off-by: Arvindh <arvindh91@gmail.com> * add group incomplete repo implementation Signed-off-by: Arvindh <arvindh91@gmail.com> * groups for users Signed-off-by: Arvindh <arvindh91@gmail.com> * groups for users Signed-off-by: Arvindh <arvindh91@gmail.com> * groups for users Signed-off-by: Arvindh <arvindh91@gmail.com> * groups for users Signed-off-by: Arvindh <arvindh91@gmail.com> --------- Signed-off-by: Arvindh <arvindh91@gmail.com> Move coap mqtt and ws policies to spicedb (#16) Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> Remove old policies Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> NOISSUE - Things authorize to return thingID (#18) This commit modifies the authorize endpoint to the grpc endpoint to return thingID. The authorize endpoint allows adapters to get the publisher of the message. Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> Add Groups to users service (#17) * add assign and unassign to group Signed-off-by: Arvindh <arvindh91@gmail.com> * add group incomplete repo implementation Signed-off-by: Arvindh <arvindh91@gmail.com> * groups for users Signed-off-by: Arvindh <arvindh91@gmail.com> * groups for users Signed-off-by: Arvindh <arvindh91@gmail.com> * groups for users Signed-off-by: Arvindh <arvindh91@gmail.com> * groups for users Signed-off-by: Arvindh <arvindh91@gmail.com> * groups for users stable 1 Signed-off-by: Arvindh <arvindh91@gmail.com> * groups for users stable 2 Signed-off-by: Arvindh <arvindh91@gmail.com> * groups for users & things Signed-off-by: Arvindh <arvindh91@gmail.com> * Amend signature Signed-off-by: Arvindh <arvindh91@gmail.com> * fix merge error Signed-off-by: Arvindh <arvindh91@gmail.com> --------- Signed-off-by: Arvindh <arvindh91@gmail.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * NOISSUE - Fix es code (#21) Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * NOISSUE - Fix Bugs (#20) * fix bugs Signed-off-by: Arvindh <arvindh91@gmail.com> * fix bugs Signed-off-by: Arvindh <arvindh91@gmail.com> --------- Signed-off-by: Arvindh <arvindh91@gmail.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * NOISSUE - Test e2e (#19) * fix: connect method Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> * fix: e2e Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> * fix changes in sdk and e2e Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> * feat(docker): remove unnecessary port mapping Remove the port mapping for MQTT broker in the docker-compose.yml file. Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> * Enable group listing Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> * feat(responses): update ChannelsPage struct The ChannelsPage struct in the responses.go file has been updated. The "Channels" field has been renamed to "Groups" to provide more accurate naming. This change ensures consistency and clarity in the codebase. Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> * feat(things): add UpdateClientSecret method Add the UpdateClientSecret method to the things service. This method allows updating the client secret for a specific client identified by the provided token, id, and key parameters. Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> --------- Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * Use smaller buffers for gRPC Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * Clean up tests (#22) Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * Add Connect Disconnect endpoints (#23) * fix bugs Signed-off-by: Arvindh <arvindh91@gmail.com> * fix bugs Signed-off-by: Arvindh <arvindh91@gmail.com> * fix list of things in a channel and Add connect disconnect endpoint Signed-off-by: Arvindh <arvindh91@gmail.com> * fix list of things in a channel and Add connect disconnect endpoint Signed-off-by: Arvindh <arvindh91@gmail.com> --------- Signed-off-by: Arvindh <arvindh91@gmail.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * Add: Things share with users (#25) * fix list of things in a channel and Add connect disconnect endpoint Signed-off-by: Arvindh <arvindh91@gmail.com> * add: things share with other users Signed-off-by: Arvindh <arvindh91@gmail.com> --------- Signed-off-by: Arvindh <arvindh91@gmail.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * NOISSUE - Rename gRPC Services (#24) * Rename things and users auth service Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> * docs: add authorization docs for gRPC services Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> * Rename things and users grpc services Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> * Remove mainflux.env package Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> --------- Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * Add: Listing of things, channels, groups, users (#26) * add: listing of channels, users, groups, things Signed-off-by: Arvindh <arvindh91@gmail.com> * add: listing of channels, users, groups, things Signed-off-by: Arvindh <arvindh91@gmail.com> * add: listing of channels, users, groups, things Signed-off-by: Arvindh <arvindh91@gmail.com> * add: listing of channels, users, groups, things Signed-off-by: Arvindh <arvindh91@gmail.com> --------- Signed-off-by: Arvindh <arvindh91@gmail.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * NOISSUE - Clean Up Users (#27) * feat(groups): rename redis package to events - Renamed the `redis` package to `events` in the `internal/groups` directory. - Updated the file paths and names accordingly. - This change reflects the more accurate purpose of the package and improves code organization. Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> * feat(auth): Modify identity method Change request and response of identity method Add accessToken and refreshToken to Token response Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> * clean up users, remove dead code Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> * feat(users): add unit tests for user service This commit adds unit tests for the user service in the `users` package. The tests cover various scenarios and ensure the correct behavior of the service. Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> --------- Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * Add: List of user groups & removed repeating code in groups (#29) * removed repeating code in list groups Signed-off-by: Arvindh <arvindh91@gmail.com> * add: list of user group Signed-off-by: Arvindh <arvindh91@gmail.com> * fix: otel handler operator name for endpoints Signed-off-by: Arvindh <arvindh91@gmail.com> --------- Signed-off-by: Arvindh <arvindh91@gmail.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * NOISSUE - Clean Up Things Service (#28) * Rework things service Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> * add tests Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> --------- Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * NOISSUE - Clean Up Auth Service (#30) * clean up auth service Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> * feat(auth): remove unused import Remove the unused import of `emptypb` in `auth.pb.go`. This import is not being used in the codebase and can be safely removed. Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> --------- Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * NOISSUE - Update API docs (#31) Signed-off-by: rodneyosodo <blackd0t@protonmail.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * Remove TODO comments and cleanup the code Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> * Update dependenices Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> --------- Signed-off-by: Arvindh <arvindh91@gmail.com> Signed-off-by: dusanb94 <dusan.borovcanin@mainflux.com> Signed-off-by: Rodney Osodo <28790446+rodneyosodo@users.noreply.github.com> Signed-off-by: rodneyosodo <blackd0t@protonmail.com> Co-authored-by: b1ackd0t <28790446+rodneyosodo@users.noreply.github.com> Co-authored-by: Arvindh <30824765+arvindh123@users.noreply.github.com>
434 lines
14 KiB
Go
434 lines
14 KiB
Go
package brotli
|
|
|
|
import "math"
|
|
|
|
/* Copyright 2013 Google Inc. All Rights Reserved.
|
|
|
|
Distributed under MIT license.
|
|
See file LICENSE for detail or copy at https://opensource.org/licenses/MIT
|
|
*/
|
|
|
|
func initialEntropyCodesDistance(data []uint16, length uint, stride uint, num_histograms uint, histograms []histogramDistance) {
|
|
var seed uint32 = 7
|
|
var block_length uint = length / num_histograms
|
|
var i uint
|
|
clearHistogramsDistance(histograms, num_histograms)
|
|
for i = 0; i < num_histograms; i++ {
|
|
var pos uint = length * i / num_histograms
|
|
if i != 0 {
|
|
pos += uint(myRand(&seed) % uint32(block_length))
|
|
}
|
|
|
|
if pos+stride >= length {
|
|
pos = length - stride - 1
|
|
}
|
|
|
|
histogramAddVectorDistance(&histograms[i], data[pos:], stride)
|
|
}
|
|
}
|
|
|
|
func randomSampleDistance(seed *uint32, data []uint16, length uint, stride uint, sample *histogramDistance) {
|
|
var pos uint = 0
|
|
if stride >= length {
|
|
stride = length
|
|
} else {
|
|
pos = uint(myRand(seed) % uint32(length-stride+1))
|
|
}
|
|
|
|
histogramAddVectorDistance(sample, data[pos:], stride)
|
|
}
|
|
|
|
func refineEntropyCodesDistance(data []uint16, length uint, stride uint, num_histograms uint, histograms []histogramDistance) {
|
|
var iters uint = kIterMulForRefining*length/stride + kMinItersForRefining
|
|
var seed uint32 = 7
|
|
var iter uint
|
|
iters = ((iters + num_histograms - 1) / num_histograms) * num_histograms
|
|
for iter = 0; iter < iters; iter++ {
|
|
var sample histogramDistance
|
|
histogramClearDistance(&sample)
|
|
randomSampleDistance(&seed, data, length, stride, &sample)
|
|
histogramAddHistogramDistance(&histograms[iter%num_histograms], &sample)
|
|
}
|
|
}
|
|
|
|
/* Assigns a block id from the range [0, num_histograms) to each data element
|
|
in data[0..length) and fills in block_id[0..length) with the assigned values.
|
|
Returns the number of blocks, i.e. one plus the number of block switches. */
|
|
func findBlocksDistance(data []uint16, length uint, block_switch_bitcost float64, num_histograms uint, histograms []histogramDistance, insert_cost []float64, cost []float64, switch_signal []byte, block_id []byte) uint {
|
|
var data_size uint = histogramDataSizeDistance()
|
|
var bitmaplen uint = (num_histograms + 7) >> 3
|
|
var num_blocks uint = 1
|
|
var i uint
|
|
var j uint
|
|
assert(num_histograms <= 256)
|
|
if num_histograms <= 1 {
|
|
for i = 0; i < length; i++ {
|
|
block_id[i] = 0
|
|
}
|
|
|
|
return 1
|
|
}
|
|
|
|
for i := 0; i < int(data_size*num_histograms); i++ {
|
|
insert_cost[i] = 0
|
|
}
|
|
for i = 0; i < num_histograms; i++ {
|
|
insert_cost[i] = fastLog2(uint(uint32(histograms[i].total_count_)))
|
|
}
|
|
|
|
for i = data_size; i != 0; {
|
|
i--
|
|
for j = 0; j < num_histograms; j++ {
|
|
insert_cost[i*num_histograms+j] = insert_cost[j] - bitCost(uint(histograms[j].data_[i]))
|
|
}
|
|
}
|
|
|
|
for i := 0; i < int(num_histograms); i++ {
|
|
cost[i] = 0
|
|
}
|
|
for i := 0; i < int(length*bitmaplen); i++ {
|
|
switch_signal[i] = 0
|
|
}
|
|
|
|
/* After each iteration of this loop, cost[k] will contain the difference
|
|
between the minimum cost of arriving at the current byte position using
|
|
entropy code k, and the minimum cost of arriving at the current byte
|
|
position. This difference is capped at the block switch cost, and if it
|
|
reaches block switch cost, it means that when we trace back from the last
|
|
position, we need to switch here. */
|
|
for i = 0; i < length; i++ {
|
|
var byte_ix uint = i
|
|
var ix uint = byte_ix * bitmaplen
|
|
var insert_cost_ix uint = uint(data[byte_ix]) * num_histograms
|
|
var min_cost float64 = 1e99
|
|
var block_switch_cost float64 = block_switch_bitcost
|
|
var k uint
|
|
for k = 0; k < num_histograms; k++ {
|
|
/* We are coding the symbol in data[byte_ix] with entropy code k. */
|
|
cost[k] += insert_cost[insert_cost_ix+k]
|
|
|
|
if cost[k] < min_cost {
|
|
min_cost = cost[k]
|
|
block_id[byte_ix] = byte(k)
|
|
}
|
|
}
|
|
|
|
/* More blocks for the beginning. */
|
|
if byte_ix < 2000 {
|
|
block_switch_cost *= 0.77 + 0.07*float64(byte_ix)/2000
|
|
}
|
|
|
|
for k = 0; k < num_histograms; k++ {
|
|
cost[k] -= min_cost
|
|
if cost[k] >= block_switch_cost {
|
|
var mask byte = byte(1 << (k & 7))
|
|
cost[k] = block_switch_cost
|
|
assert(k>>3 < bitmaplen)
|
|
switch_signal[ix+(k>>3)] |= mask
|
|
/* Trace back from the last position and switch at the marked places. */
|
|
}
|
|
}
|
|
}
|
|
{
|
|
var byte_ix uint = length - 1
|
|
var ix uint = byte_ix * bitmaplen
|
|
var cur_id byte = block_id[byte_ix]
|
|
for byte_ix > 0 {
|
|
var mask byte = byte(1 << (cur_id & 7))
|
|
assert(uint(cur_id)>>3 < bitmaplen)
|
|
byte_ix--
|
|
ix -= bitmaplen
|
|
if switch_signal[ix+uint(cur_id>>3)]&mask != 0 {
|
|
if cur_id != block_id[byte_ix] {
|
|
cur_id = block_id[byte_ix]
|
|
num_blocks++
|
|
}
|
|
}
|
|
|
|
block_id[byte_ix] = cur_id
|
|
}
|
|
}
|
|
|
|
return num_blocks
|
|
}
|
|
|
|
var remapBlockIdsDistance_kInvalidId uint16 = 256
|
|
|
|
func remapBlockIdsDistance(block_ids []byte, length uint, new_id []uint16, num_histograms uint) uint {
|
|
var next_id uint16 = 0
|
|
var i uint
|
|
for i = 0; i < num_histograms; i++ {
|
|
new_id[i] = remapBlockIdsDistance_kInvalidId
|
|
}
|
|
|
|
for i = 0; i < length; i++ {
|
|
assert(uint(block_ids[i]) < num_histograms)
|
|
if new_id[block_ids[i]] == remapBlockIdsDistance_kInvalidId {
|
|
new_id[block_ids[i]] = next_id
|
|
next_id++
|
|
}
|
|
}
|
|
|
|
for i = 0; i < length; i++ {
|
|
block_ids[i] = byte(new_id[block_ids[i]])
|
|
assert(uint(block_ids[i]) < num_histograms)
|
|
}
|
|
|
|
assert(uint(next_id) <= num_histograms)
|
|
return uint(next_id)
|
|
}
|
|
|
|
func buildBlockHistogramsDistance(data []uint16, length uint, block_ids []byte, num_histograms uint, histograms []histogramDistance) {
|
|
var i uint
|
|
clearHistogramsDistance(histograms, num_histograms)
|
|
for i = 0; i < length; i++ {
|
|
histogramAddDistance(&histograms[block_ids[i]], uint(data[i]))
|
|
}
|
|
}
|
|
|
|
var clusterBlocksDistance_kInvalidIndex uint32 = math.MaxUint32
|
|
|
|
func clusterBlocksDistance(data []uint16, length uint, num_blocks uint, block_ids []byte, split *blockSplit) {
|
|
var histogram_symbols []uint32 = make([]uint32, num_blocks)
|
|
var block_lengths []uint32 = make([]uint32, num_blocks)
|
|
var expected_num_clusters uint = clustersPerBatch * (num_blocks + histogramsPerBatch - 1) / histogramsPerBatch
|
|
var all_histograms_size uint = 0
|
|
var all_histograms_capacity uint = expected_num_clusters
|
|
var all_histograms []histogramDistance = make([]histogramDistance, all_histograms_capacity)
|
|
var cluster_size_size uint = 0
|
|
var cluster_size_capacity uint = expected_num_clusters
|
|
var cluster_size []uint32 = make([]uint32, cluster_size_capacity)
|
|
var num_clusters uint = 0
|
|
var histograms []histogramDistance = make([]histogramDistance, brotli_min_size_t(num_blocks, histogramsPerBatch))
|
|
var max_num_pairs uint = histogramsPerBatch * histogramsPerBatch / 2
|
|
var pairs_capacity uint = max_num_pairs + 1
|
|
var pairs []histogramPair = make([]histogramPair, pairs_capacity)
|
|
var pos uint = 0
|
|
var clusters []uint32
|
|
var num_final_clusters uint
|
|
var new_index []uint32
|
|
var i uint
|
|
var sizes = [histogramsPerBatch]uint32{0}
|
|
var new_clusters = [histogramsPerBatch]uint32{0}
|
|
var symbols = [histogramsPerBatch]uint32{0}
|
|
var remap = [histogramsPerBatch]uint32{0}
|
|
|
|
for i := 0; i < int(num_blocks); i++ {
|
|
block_lengths[i] = 0
|
|
}
|
|
{
|
|
var block_idx uint = 0
|
|
for i = 0; i < length; i++ {
|
|
assert(block_idx < num_blocks)
|
|
block_lengths[block_idx]++
|
|
if i+1 == length || block_ids[i] != block_ids[i+1] {
|
|
block_idx++
|
|
}
|
|
}
|
|
|
|
assert(block_idx == num_blocks)
|
|
}
|
|
|
|
for i = 0; i < num_blocks; i += histogramsPerBatch {
|
|
var num_to_combine uint = brotli_min_size_t(num_blocks-i, histogramsPerBatch)
|
|
var num_new_clusters uint
|
|
var j uint
|
|
for j = 0; j < num_to_combine; j++ {
|
|
var k uint
|
|
histogramClearDistance(&histograms[j])
|
|
for k = 0; uint32(k) < block_lengths[i+j]; k++ {
|
|
histogramAddDistance(&histograms[j], uint(data[pos]))
|
|
pos++
|
|
}
|
|
|
|
histograms[j].bit_cost_ = populationCostDistance(&histograms[j])
|
|
new_clusters[j] = uint32(j)
|
|
symbols[j] = uint32(j)
|
|
sizes[j] = 1
|
|
}
|
|
|
|
num_new_clusters = histogramCombineDistance(histograms, sizes[:], symbols[:], new_clusters[:], []histogramPair(pairs), num_to_combine, num_to_combine, histogramsPerBatch, max_num_pairs)
|
|
if all_histograms_capacity < (all_histograms_size + num_new_clusters) {
|
|
var _new_size uint
|
|
if all_histograms_capacity == 0 {
|
|
_new_size = all_histograms_size + num_new_clusters
|
|
} else {
|
|
_new_size = all_histograms_capacity
|
|
}
|
|
var new_array []histogramDistance
|
|
for _new_size < (all_histograms_size + num_new_clusters) {
|
|
_new_size *= 2
|
|
}
|
|
new_array = make([]histogramDistance, _new_size)
|
|
if all_histograms_capacity != 0 {
|
|
copy(new_array, all_histograms[:all_histograms_capacity])
|
|
}
|
|
|
|
all_histograms = new_array
|
|
all_histograms_capacity = _new_size
|
|
}
|
|
|
|
brotli_ensure_capacity_uint32_t(&cluster_size, &cluster_size_capacity, cluster_size_size+num_new_clusters)
|
|
for j = 0; j < num_new_clusters; j++ {
|
|
all_histograms[all_histograms_size] = histograms[new_clusters[j]]
|
|
all_histograms_size++
|
|
cluster_size[cluster_size_size] = sizes[new_clusters[j]]
|
|
cluster_size_size++
|
|
remap[new_clusters[j]] = uint32(j)
|
|
}
|
|
|
|
for j = 0; j < num_to_combine; j++ {
|
|
histogram_symbols[i+j] = uint32(num_clusters) + remap[symbols[j]]
|
|
}
|
|
|
|
num_clusters += num_new_clusters
|
|
assert(num_clusters == cluster_size_size)
|
|
assert(num_clusters == all_histograms_size)
|
|
}
|
|
|
|
histograms = nil
|
|
|
|
max_num_pairs = brotli_min_size_t(64*num_clusters, (num_clusters/2)*num_clusters)
|
|
if pairs_capacity < max_num_pairs+1 {
|
|
pairs = nil
|
|
pairs = make([]histogramPair, (max_num_pairs + 1))
|
|
}
|
|
|
|
clusters = make([]uint32, num_clusters)
|
|
for i = 0; i < num_clusters; i++ {
|
|
clusters[i] = uint32(i)
|
|
}
|
|
|
|
num_final_clusters = histogramCombineDistance(all_histograms, cluster_size, histogram_symbols, clusters, pairs, num_clusters, num_blocks, maxNumberOfBlockTypes, max_num_pairs)
|
|
pairs = nil
|
|
cluster_size = nil
|
|
|
|
new_index = make([]uint32, num_clusters)
|
|
for i = 0; i < num_clusters; i++ {
|
|
new_index[i] = clusterBlocksDistance_kInvalidIndex
|
|
}
|
|
pos = 0
|
|
{
|
|
var next_index uint32 = 0
|
|
for i = 0; i < num_blocks; i++ {
|
|
var histo histogramDistance
|
|
var j uint
|
|
var best_out uint32
|
|
var best_bits float64
|
|
histogramClearDistance(&histo)
|
|
for j = 0; uint32(j) < block_lengths[i]; j++ {
|
|
histogramAddDistance(&histo, uint(data[pos]))
|
|
pos++
|
|
}
|
|
|
|
if i == 0 {
|
|
best_out = histogram_symbols[0]
|
|
} else {
|
|
best_out = histogram_symbols[i-1]
|
|
}
|
|
best_bits = histogramBitCostDistanceDistance(&histo, &all_histograms[best_out])
|
|
for j = 0; j < num_final_clusters; j++ {
|
|
var cur_bits float64 = histogramBitCostDistanceDistance(&histo, &all_histograms[clusters[j]])
|
|
if cur_bits < best_bits {
|
|
best_bits = cur_bits
|
|
best_out = clusters[j]
|
|
}
|
|
}
|
|
|
|
histogram_symbols[i] = best_out
|
|
if new_index[best_out] == clusterBlocksDistance_kInvalidIndex {
|
|
new_index[best_out] = next_index
|
|
next_index++
|
|
}
|
|
}
|
|
}
|
|
|
|
clusters = nil
|
|
all_histograms = nil
|
|
brotli_ensure_capacity_uint8_t(&split.types, &split.types_alloc_size, num_blocks)
|
|
brotli_ensure_capacity_uint32_t(&split.lengths, &split.lengths_alloc_size, num_blocks)
|
|
{
|
|
var cur_length uint32 = 0
|
|
var block_idx uint = 0
|
|
var max_type byte = 0
|
|
for i = 0; i < num_blocks; i++ {
|
|
cur_length += block_lengths[i]
|
|
if i+1 == num_blocks || histogram_symbols[i] != histogram_symbols[i+1] {
|
|
var id byte = byte(new_index[histogram_symbols[i]])
|
|
split.types[block_idx] = id
|
|
split.lengths[block_idx] = cur_length
|
|
max_type = brotli_max_uint8_t(max_type, id)
|
|
cur_length = 0
|
|
block_idx++
|
|
}
|
|
}
|
|
|
|
split.num_blocks = block_idx
|
|
split.num_types = uint(max_type) + 1
|
|
}
|
|
|
|
new_index = nil
|
|
block_lengths = nil
|
|
histogram_symbols = nil
|
|
}
|
|
|
|
func splitByteVectorDistance(data []uint16, length uint, literals_per_histogram uint, max_histograms uint, sampling_stride_length uint, block_switch_cost float64, params *encoderParams, split *blockSplit) {
|
|
var data_size uint = histogramDataSizeDistance()
|
|
var num_histograms uint = length/literals_per_histogram + 1
|
|
var histograms []histogramDistance
|
|
if num_histograms > max_histograms {
|
|
num_histograms = max_histograms
|
|
}
|
|
|
|
if length == 0 {
|
|
split.num_types = 1
|
|
return
|
|
} else if length < kMinLengthForBlockSplitting {
|
|
brotli_ensure_capacity_uint8_t(&split.types, &split.types_alloc_size, split.num_blocks+1)
|
|
brotli_ensure_capacity_uint32_t(&split.lengths, &split.lengths_alloc_size, split.num_blocks+1)
|
|
split.num_types = 1
|
|
split.types[split.num_blocks] = 0
|
|
split.lengths[split.num_blocks] = uint32(length)
|
|
split.num_blocks++
|
|
return
|
|
}
|
|
|
|
histograms = make([]histogramDistance, num_histograms)
|
|
|
|
/* Find good entropy codes. */
|
|
initialEntropyCodesDistance(data, length, sampling_stride_length, num_histograms, histograms)
|
|
|
|
refineEntropyCodesDistance(data, length, sampling_stride_length, num_histograms, histograms)
|
|
{
|
|
var block_ids []byte = make([]byte, length)
|
|
var num_blocks uint = 0
|
|
var bitmaplen uint = (num_histograms + 7) >> 3
|
|
var insert_cost []float64 = make([]float64, (data_size * num_histograms))
|
|
var cost []float64 = make([]float64, num_histograms)
|
|
var switch_signal []byte = make([]byte, (length * bitmaplen))
|
|
var new_id []uint16 = make([]uint16, num_histograms)
|
|
var iters uint
|
|
if params.quality < hqZopflificationQuality {
|
|
iters = 3
|
|
} else {
|
|
iters = 10
|
|
}
|
|
/* Find a good path through literals with the good entropy codes. */
|
|
|
|
var i uint
|
|
for i = 0; i < iters; i++ {
|
|
num_blocks = findBlocksDistance(data, length, block_switch_cost, num_histograms, histograms, insert_cost, cost, switch_signal, block_ids)
|
|
num_histograms = remapBlockIdsDistance(block_ids, length, new_id, num_histograms)
|
|
buildBlockHistogramsDistance(data, length, block_ids, num_histograms, histograms)
|
|
}
|
|
|
|
insert_cost = nil
|
|
cost = nil
|
|
switch_signal = nil
|
|
new_id = nil
|
|
histograms = nil
|
|
clusterBlocksDistance(data, length, num_blocks, block_ids, split)
|
|
block_ids = nil
|
|
}
|
|
}
|