• No results found

Convolutive Features for Storage and Transmission

N/A
N/A
Protected

Academic year: 2022

Share "Convolutive Features for Storage and Transmission"

Copied!
1
0
0

Loading.... (view fulltext now)

Full text

(1)

Convolutive Features for Storage and Transmission

Ruair´ı de Fr´ein hrdefrein@gmail.comi. KTH - Royal Institute of Technology, Sweden & Telecommunications Software and Systems Group, Ireland

Abstract

A central concern for many learning algorithms and sensing

systems is how to efficiently store what the algorithm/system has learned. Convolutive Non-negative Matrix Factorization (CNMF) finds parts-based convolutive representations of non-negative data. Convolutive extensions of NMF have not considered storage efficiency as a side constraint during the learning procedure. We contribute an algorithm, Storable NMF (SNMF), that fuses ideas from the (1) parts-based learning and (2) integer sequence

compression literature. SNMF enjoys the merits of both

techniques: it retains the good-approximation properties of CNMF while also taking into account the size of the symbol set which is used to express the learned convolutive factors and

activations. We demonstrate that SNMF yields a compression

ratio ranging from 10:1 up to 20:1, which gives rise to a similar bandwidth and storage saving for networked sensors.

Trick: SNMF achieves these improved compression ratios,

without incurring a significant loss of accuracy, by embedding an off-the-shelf compression algorithm in the CNMF updates so that quantization function updates are interleaved with CNMF’s up- date rules.

Introduction

• Non-negative Matrix Factorization (NMF): bi-linear decomposition of multivariate data [1, 2].

• NMF factorizes V R+M×N into the product of two matrices W R+M×R (basis) and H RR+×N (activations), yielding the estimate ˜V = W H. R < M, N.

NMF is a “parts-based” approach [3, 4, 5].

• In the case of audio, NMF is applied to magnitude spectra [6].

Reconstruction objective:

D(V||V˜ ) =

m,n

V[m, n] log V[m, n]

V˜ [m, n] V[m, n] + V˜ [m, n]. (1)

Motivation

Learn acoustic features [7] at the edge of a network that can be transferred between (and stored on) different

sensors/networked devices.

We exploit Occam’s razor by:

• Learning low-rank factorization;

• Learning parsimonious activation functions by using a convolutive generative model;

• Expressing the factorization using a symbol set with a small car- dinality.

Convolutive NMF

• Extends NMF to time-varying objects [7] (capture features with temporal extent);

V V˜ =

T1 t

=0

Wt tH . (2)

• Activates a cascade of T basis functions;

H = H W

Tt

t

hV V˜

i

WtT1M×N, Wt = Wt

V˜ V

tT

H

1M×N tHT

(3)

• Learns T + 1 matrices W0, W1 . . . WT1, and H, where H is the average update.

Evolving Spectrogram: Signal composed of auditory objects that evolve with time; Basis functions (that evolve with time); Acti- vations of these functions [8].

This work was supported by an ELEVATE Irish Research Council International Career Devel- opment Fellowship co-funded by Marie Curie Actions award: ELEVATEPD/2014/62.

Storable NMF

• Taking a NLA at convergence to improve compressibility of

factors, introduces error indiscriminately (b·c denotes rounding and ? is the quantization step-size);

Hˆ  1

?H + 1 2



, Wˆ t



Wt 1

? + 1 2



, (4)

where ˆWt[m, r], ˆH[r, n] ∈ I.

I =

[

m,r,t

Wˆ t[m, r]

[

(

[

r,n

Hˆ [r, n] )

. (5)

Is there a better way to pick the set?

Adapt the step-size for each row of H [9, 10]

Hˆ [1, n] H[1, n]

h[1]

Hˆ [2, n] H[2, n]

h[2]

Adaptive Quantization

• Adapt the quantization function’s step-size for each column of Wt and each row H.

Wˆ t[:, r] ←  Wt[:, r]

w[r] + 1 2



, r, t Hˆ [r, :] ←  H[r, :]

h[r] + 1 2



, r. (6)

• Initially I is a subset of I R+; At convergence, a subset of the positive integers.

• The cardinality of I, is generally minimized as I is indirectly learned by the factorization.

Definition 1 SNMF factorizes the matrix V R+M×N into the product of two element-wise integer matrices ˆWt IM×R and ˆH IR×N and two diagonal matrices diag(w) and diag(h),

V˜ =

T1 t

=0

Wˆ tdiag(w)diag(h)

t

Hˆ , where w, h RR+×1. (7)

• Quantization step-sizes, w[r], h[r] ∈ R+, though real-valued, are few, and are adapted to reduce the cardinality of I. The majority of elements of the factors are integer, Wt[m, r], tH [r, n] ∈ I.

Quantization Step-size Updates: The quantization step-sizes

h[r] and w[r], in the convolutional mixing model in (Eqn. 7) produce NN updates if they are updated using

h[r] ← h[r] 1TM

"

V Vˆ

"

Tt=01Wˆ t[:, r]w[r]

t

Hˆ [r, :]

##

1N

1TM

"

Tt=01Wˆ t[:, r]w[r]

t

Hˆ [r, :]

# 1N

, (8)

w[r] ← w[r] 1TM

"

V Vˆ

"

Tt=01Wˆ t[:, r]h[r]

t

Hˆ [r, :]

##

1N

1TM

"

Tt=01Wˆ t[:, r]h[r]

t

Hˆ [r, :]

# 1N

. (9)

NLA b·c & Monotonic Convergence

• Quantization of elements of ˆWt may cause the KLD to increase;

• Generally, W[r]  Wˆ t[:, r] (for the important (big) entries);

• We have freedom in how we apply element-wise rounding; we desire monotonic convergence. Heuristic: we randomly

substitute in a number of new quantization functions until KLD is minimized.

Choose between

 Wt[m, r]

W[r] + 1 2



m,r,t

, and  Wt[m, r]

w[r] 1 2



m,r,t

(10) with probability p.

SNMF-parsing-encoding scheme

• Concatenate the columns of ˆW and take transpose to generate a sequence

Wt = [Wˆ t[:, 1]T|Wˆ t[:, 2]T| . . . |Wˆ t[:, R]T]. (11)

• Concatenate the rows of ˆH to generate,

H = [Hˆ [1, :]|Hˆ [2, :]| . . . |Hˆ [R, :]]. (12)

• Generate the sequence of integers,

S = M|R|N|W|H|w|h. (13)

• Works well because gzip adapts to changes in the integer

sequence. Different columns and rows of ˆW & ˆH have different stats.

1 1 3 1 1 3 1 2 3 1 2 1 1 1 3 7→ , 0, 1)

Look ahead

1 1 3 1 1 3 1 2 3 1 2 1 1 1 3 7→ (1, 1, 3)

Look ahead

1 1 3 1 1 3 1 2 3 1 2 1 1 1 3 7→ (3, 4, 2)

Look ahead Dictionary

1 1 3 1 1 3 1 2 3 1 2 1 1 1 3 7→ (3, 3, 1)

Look ahead Dictionary

1 1 3 1 1 3 1 2 3 1 2 1 1 1 3 7→ (1, 2, 3)

Look ahead Dictionary

2015 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this materialc for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE. For more details: Ruair´ı de Fr´ein, ”Learning Convolutive Features for Storage and Transmission between Networked Sensors”, 2015 International Joint Conference on Neural Networks, to appear.

Results

Hypothesis: Given that (C)NMF is non-unique and non-exact, but typically yields a satisfactory decomposition, we have the

freedom to choose an arbitrary satisfactory lossy factorization with respect to the distortion constraint of choice. Can SNMF learn a factorization which is: (1) within 1-3dB of the expected CNMF Signal-to-Noise-Ratio and (2) amenable to efficient coding?

SNMF Applied to Synthetic Data-set:

We report average performance over 100 trials. Each trial was run for 100 iterations. R = 2. FFT = 1024. Overlap = 512. SNR is improved by increasing R.

• Noise-Burst Data-set: Performance comparison NMF vs SNMF (T = 1).

Algorithm NMF SNMF (T = 1)

No. unique elements of W/ ˆW 1026 138

No. unique elements of H/ ˆH 874 19

File-size (gzip / not gzipped) in bytes 17704/35839 1659/35890

Compression ratio 10:1

SNR (dB) 6.58dB 6.6dB

• Evolving Spectrogram: Performance comparison CNMF vs SNMF (T = 32).

Algorithm CNMF SNMF (T = 32)

No. unique elements of W/ ˆW 32832 155 No. unique elements of H/ ˆH 436 21

File-size (gzip/ not gzipped) 300239/625609 13841/674142

Compression ratio 20:1

SNR (dB) 12.5dB 11.56dB

Noise-Burst Data: Histograms of the elements of SNMF’s (T = 1) Wˆ vs NMF’s W.

SNMF (above); NMF (below).

Noise-Burst Data.

• The number of bins of the histograms associated with SNMF (T = 1) is smaller than for NMF.

• NMF’s basis function elements only have one element per bin.

SNMF Applied to TIMIT Database:

CNMF SNMF CNMF SNMF

Algorithm (T = 1) (T = 1) (T = 5) (T = 5) No. unique elements W/ ˆW 10260 1133 51299 374 No. unique elements H/ ˆH 1820 518 1818 54 File-size gzip bytes 112278 11442 489705 46389 File-size not gzipped bytes 232444 226119 1016321 1035544

Compression ratio 9.8:1 10.6:1

SNR (dB) 11.61 11.19 25.4 24.14

Algorithm (T = 3) (T = 3) (T = 7) (T = 7) No. unique elements W/ ˆW 30780 341 71820 421 No. unique elements H/ ˆH 1820 86 1801 47 File-size gzip bytes 305111 30243 682557 63990 File-size not gzipped bytes 632236 633672 1417852 1438136

Compression ratio 10.1:1 10.67:1

SNR (dB) 15.63 15.83 29.28 26.5

• Effect of T on CNMF versus SNMF (1 T 7);

• Each trial run for 500 iterations; R = 20 is fixed;

• Set I of SNMF is smaller than CNMF;

• Compressibility of SNMF is greater than CNMF;

• Compressibility is not affected by an increase in T;

• Error introduced by SNMF is reasonable relative to CNMF.

Conclusions

Element-wise integer constraints are introduced and

compression is improved. SNMF combined with a universal loss-less coding scheme is portable and efficiently stored.

Error introduced by integerizing the factors is folded into the CNMF updates. A 10:1 or 20:1 compression ratio of the results of a (C)NMF is possible if the factors are restricted to a finite integer symbol set. This only incurs a distortion cost of 1-3dB.

References

[1] D.D. Lee and H.S. Seung, “Algorithms for Nonnegative Matrix Factorization,” NIPS, vol.

13, pp. 556–62, 2001.

[2] R. de Fr´ein, K. Drakakis, and S. Rickard, “Portfolio diversification using subspace factori- sations,” in IEEE Inf. Sc. Sys. 42nd Conf., 2008, pp. 1075–80.

[3] D. Donoho and V. Stodden, “When does Nonnegative Matrix Factorization give a correct decomposition into parts?,” NIPS, vol. 16, 2004.

[4] Ruair´ı de Fr´ein, “Formal concept analysis via atomic priming,” in Formal Concept Analysis, Peggy Cellier, Felix Distel, and Bernhard Ganter, Eds., vol. 7880 of Lecture Notes in Computer Science, pp. 92–108. Springer Berlin Heidelberg, 2013.

[5] Ruair´ı de Fr´ein, “Ghostbusters: A parts-based NMF algorithm,” in Signals and Systems Conference (ISSC 2013), 24th IET Irish, Jun. 2013, pp. 1–8.

[6] R.K. Potter, G.A. Kopp, and H.C. Green, “Visible speech,” 1947.

[7] P. Smaragdis, “Nonnegative matrix factor deconvolution; extraction of multiple sound sources from monophonic inputs,” Int. Conf. Independent Component Analysis and Blind Sig- nal Separation, pp. 494–9, 2004.

[8] P. D. O’Grady and B. A. Pearlmutter, “Discovering speech phones using convolutive non- negative matrix factorisation with a sparseness constraint,” in Neurocomp., 2008, pp. 88–

101.

[9] R. de Fr´ein, “Learning and Storing the Parts of Objects: IMF,” in IEEE Int. Workshop on Machine Learning for Sig. Proc., Sep. 2014, pp. 1–6.

[10] R. de Frein, “Quantized nonnegative matrix factorization,” in Digital Signal Processing (DSP), 2014 19th International Conference on, Aug 2014, pp. 377–382.

1

References

Related documents

Given several trips that share a road segment, we can calculate FPCs for those trips, and then perform statistical inference on any other variable, such as traffic based on time of

The purpose of this thesis is to provide the overview of different methods to cal- culate the fault distance and to distinguish two kinds of transients faults, single

We combine several approaches into an ensemble, based on spelling error features, a simple neural net- work using word representations, a deep residual network using word and

By linearly interpolating each movie segment from the VQM scores and file sizes of this segment in two consecutive resolutions that are closest to the desired VQM threshold, we

The object’s dimension and orientation are estimated using rectangle fitting. The rectangle fitting problem is expressed using non-convex optimization, least squares optimization

From the landfills leaching water with high PFAS concentration is released to the system but the small flow of the leaching water will be diluted so the impact on the sampling

Återigen gör underrätterna en bedömning av huruvida gärningsmannen haft eventuellt uppsåt att skada eller inte och konstaterar att med hänsyn till den trängda situationen i

Intressant nog går denna förändring hand i hand med inte bara elektrifiering och ett avfärdande av den äldre civilisationskritiken, utan också med en förändring av placering- en